var/home/core/zuul-output/0000755000175000017500000000000015156366001014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015156372441015501 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000245117415156372266020301 0ustar corecoreikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9GfͅR~i.߷;U/;Yw?.y7W޾n^X/ixK|1Ool_~yyiw|zxV^֯v5gCh31 )Kh3i J1hG{aD4iӌçN/e] o;iF]u54!h/9Y@$9GAOI=2,!N{\00{B"唄(".V.U) _.f*g,Z0>?<;~9.뙘 vKAb;-$JRPţ*描Լf^`iwoW~SL2uQO)qai]>yE*,?k 9Z29}}(4ҲIFyG -^W6yY<*uvf d |TRZ;j?| |!I糓 sw`{s0Aȶ9W E%*mG:tëoG(;h0!}qfJz硂Ϧ4Ck9]٣Z%T%x~5r.N`$g`Խ!:*Wni|QXj0NbYe獸]fNdƭwq <ć;_ʧNs9[(=!@Q,}s=LN YlYd'Z;o.K'_-הp|A*Z*}QJ0SqAYE0i5P-$̿<_d^"]}Z|-5rC wjof'(%*݅^J">CMMQQ؏*ΧL ߁NPi?$;g&立q^-:SӨwٗ*AqΝ6 lv?Qw):yt٘8c'8Ai؋%\78:ZBBk`E\Ƹ#¿Øp*vxyPLSMY 9J}t/A`*t) O5]/* @.yhi-cS4 6"KaFٗt<>vRڡc0SAA\c}or|MKrO] g"tta[I!;c%6$V<[+*J:AI \:-rR b B"~?4 W4B3lLRD|@Kfځ9g ? j럚Sř>]uw`@}-{C):fUr6v`mSΟ1c/n?!'Y|7#RI)X)yCBoX^P\Ja 79clw/H tBFKskޒ1,%$BվCh,xɦSBZa;0RZ+ 9O5KiPc7C>a~W;D=;y|AAY'"葋_d$Ə{(he NSfX1982TH#D֪v3l"<, { Tms'oI&'Adp]{1DL^5"Ϧޙ`F}W5XDV7V5EE9esYYfiMOV i/ f>3VQ 7,oTW⇊AqO:rƭĘ DuZ^ To3dEN/} fI+?|Uz5SUZa{P,97óI,Q{eNFV+(hʺb ״ʻʞX6ýcsT z`q 0C?41- _n^ylSO2|'P'BOTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8c3ilLJ!Ip,2(( *%KGj   %*e5-wFp"a~fzqu6tY,d,`!qIv꜒"T[1!I!NwL}\|}.b3oXR\(L _nJBR_v'5n]FhNU˿oۂ6C9C7sn,kje*;iΓA7,Q)-,=1A sK|ۜLɽy]ʸEO<-YEqKzϢ \{>dDLF amKGm+`VLJsC>?5rk{-3Ss`y_C}Q v,{*)ߎ% qƦat:D=uNvdߋ{Ny[$ {ɴ6hOI']dC5`t9:GO: FmlN*:g^;T^B0$B%C6Θ%|5u=kkN2{'FEc* A>{avdt)8|mg定TN7,TEXt+`F P |ɧ<Ғ8_iqE b}$B#fethBE;1"l r  B+R6Qp%;R8P󦟶Ub-L::;Ⱦ7,VW.JE:PgXoΰUv:ΰdɆΰ (ΰ0eTUgXun[g, ֶ Jy#cAz~M5Ys_T=D*A+6J'2I+6֗%)FV3_F;TM\jP' HT؋ Bvaijz,lzvK2Zu8C1}PcIr.e'.I(NJKۨzHJodo\4"+(Nck!a}P `x ăiP=!X;__ZNۈO{n[>77ĐrX(K&Y5+$wL#ɽ 4d-bbdAJ?:P>n^2] e}gjFX@&avF묇cTy^}m .Ŏ7Uֻ󂊹P-\!3^.Y9[XԦo Έ')Ji.VՕH4~)(kKC&;嶑, }t&&\5u17\I@ 5O? ʴ(aPqPϟ'+'h=TԫeVިO? )-1 8/%\hC(:=4< ,RmDRWfRoUJy ŗ-ܲ(4k%הrΒ]rύW -e]hx&gs7,6BxzxօoFMA['҉F=NGD4sTq1HPld=Q,DQ IJipqc2*;/!~x]y7D7@u邗`unn_ư-a9t_/.9tTo]r8-X{TMYtt =0AMUk}G9^UA,;Tt,"Dxl DfA\w; &`Ͱ٢x'H/jh7hM=~ ֟y[dI~fHIqC۶1Ik\)3 5Ķ']?SؠC"j_6Ÿ9؎]TTjm\D^x6ANbC ]tVUKe$,\ܺI `Qز@UӬ@B {~6caR!=A>\+܁<lW Gϸ}^w'̅dk  C 7fbU{3Se[s %'!?xL 2ڲ]>i+m^CM&WTj7ȗE!NC6P}H`k(FUM gul)b ;2n6'k}ˍ[`-fYX_pL +1wu(#'3"fxsuҮױdy.0]?ݽb+ uV4}rdM$ѢIA$;~Lvigu+]NC5ÿ nNჶT@~ܥ 7-mU,\rXmQALglNʆ P7k%v>"WCyVtnV K`pC?fE?~fjBwU&'ᚡilRї`m] leu]+?T4v\% ;qF0qV(]pP4W =d#t ru\M{Nj.~27)p|Vn60֭l$4԰vg`i{ 6uwŇctyX{>GXg&[ņzP8_ "J~7+0_t[%XU͍ &dtO:odtRWon%*44nc.^xt4gD638L"!}LpInTeD_1ZrbkI%8zPU:LNTPlI&N:o&2BVb+uxZ`v?7"I8hp A&?a(8E-DHa%LMg2:-ŷX(ǒ>,ݵ𴛾é5Zٵ]z"]òƓVgzEY9[Nj_vZ :jJ2^b_ F w#X6Sho禮<u8.H#',c@V8 iRX &4ڻ8zݽ.7jhvQ:H0Np: qfՋ40oW&&ף \9ys8;ӷL:@۬˨vvn/sc}2N1DDa(kx.L(f"-Da +iP^]OrwY~fwA#ٔ!:*땽Zp!{g4څZtu\1!ѨW(7qZcpL)ύ-G~^rFD+"?_h)yh=x>5ܙQ~O_e琇HBzI7*-Oi* VšPȰһ8hBőa^mX%SHR Fp)$J7A3&ojp/68uK͌iΙINmq&} O L-\ n4f/uc:7k]4p8wWLeUc.)#/udoz$} _3V6UݎvxyRC%ƚq5Щ/ۅw* CVo-1딆~ZYfJ"ou1ϵ5E bQ2mOΏ+w_eaxxOq:ym\q!<'J[FJ,4N:=6. +;$v6"I7%#CLTLyi{+ɠ^^fRa6ܮIN ޖ:DMz'rx#~w7U6=S0+ň+[Miw(W6 ]6ȧyԋ4ԙ./_A9B_-Z\PM `iĸ&^Ut (6{\٢K 5XGU/m >6JXa5FA@ q}4BooRe&#c5t'B6Ni/~?aX9QR5'%9hb,dsPn2Y??N M<0YaXJ)?ѧ| ;&kEYhjo?BOy)O˧?GϧmI C6HJ{jc kkA ~u?u7<?gd iAe1YB siҷ,vm}S|z(N%Wг5=08`S*՟݃*־%NǸ*kb05 V8[l?W]^@G:{N-i bɵFWǙ*+Ss*iނL8]iSCQ&s~In/SZ % 'I Ƿ$M6rN+LxE>^DݮEڬTk1+trǴ5RHİ{qJ\}X` >+%ni3+(0m8HЭ*zAep!*)jxG:Up~gfu#x~ .2ןGRLIۘT==!TlN3ӆv%#oV}N~ˊc,_,=COU C],Ϣa!L}sy}u\0U'&2ihbvz=.ӟk ez\ƚO; -%M>AzzGvݑT58ry\wW|~3Ԟ_f&OC"msht: rF<SYi&It1!ʐDN q$0Y&Hv]9Zq=N1/u&%].]y#z18m@n1YHR=53hHT( Q(e@-#!'^AK$wTg1!H$|HBTf̋ Y@Mwq[Fī h[W,Ê=j8&d ԋU.I{7O=%iG|xqBչ̋@1+^.r%V12, _&/j"2@+ wm 4\xNtˆ;1ditQyc,m+-!sFɸv'IJ-tH{ "KFnLRH+H6Er$igsϦ>QKwҰ]Mfj8dqV+"/fC Q`B 6כy^SL[bJgW^;zA6hrH#< 1= F8) 򃟤,ŏd7>WKĉ~b2KQdk6՛tgYͼ#$eooԦ=#&d.09DHN>AK|s:.HDŽ">#%zNEt"tLvfkB|rN`)81 &ӭsēj\4iO,H̎<ߥ諵z/f]v2 0t[U;;+8&b=zwɓJ``FiQg9XʐoHKFϗ;gQZg܉?^_ XC.l.;oX]}:>3K0R|WD\hnZm֏op};ԫ^(fL}0/E>ƥN7OQ.8[ʔh,Rt:p<0-ʁקiߟt[A3)i>3Z i򩸉*ΏlA" &:1;O]-wgϊ)hn&i'v"/ͤqr@8!̴G~7u5/>HB)iYBAXKL =Z@ >lN%hwiiUsIA8Y&=*2 5I bHb3Lh!ޒh7YJt*CyJÄFKKùMt}.l^]El>NK|//f&!B {&g\,}F)L b߀My6Õw7[{Gqzfz3_X !xJ8T<2!)^_ďǂ.\-d)Kl1헐Z1WMʜ5$)M1Lʳsw5ǫR^v|t$VȖA+Lܑ,҂+sM/ѭy)_ÕNvc*@k]ן;trȫpeoxӻo_nfz6ؘҊ?b*bj^Tc?m%3-$h`EbDC;.j0X1dR? ^}Ծե4NI ܓR{Omu/~+^K9>lIxpI"wS S 'MV+Z:H2d,P4J8 L72?og1>b$]ObsKx̊y`bE&>XYs䀚EƂ@K?n>lhTm' nܡvO+0fqf٠r,$/Zt-1-dė}2Or@3?]^ʧM <mBɃkQ }^an.Fg86}I h5&XӘ8,>b _ z>9!Z>gUŞ}xTL̵ F8ՅX/!gqwߑZȖF 3U>gCCY Hsc`% s8,A_R$קQM17h\EL#w@>omJ/ŵ_iݼGw eIJipFrO{uqy/]c 2ėi_e}L~5&lҬt񗽐0/λL[H* JzeMlTr &|R 2ӗh$cdk?vy̦7]Ạ8ph?z]W_MqKJ> QA^"nYG0_8`N 7{Puٽ/}3ymGqF8RŔ.MMWrO»HzC7ݴLLƓxxi2mW4*@`tF)Ċ+@@te\0zE|!@E " ;9Ώf3kZc7B 8yݪkIf-8>V#ہll/ؽnA(ȱbAj>C9O n6HNe">0]8@*0)QsUN8t^N+mXU q2EDö0^R) hCt{d}ܜFnԴ.2w⠪R/r| w,?VMqܙ7;qpUۚ5Tnj ۝jlN$q:w$U>tL)NC*<` `)ĉJآS2 z]gQ)Bی:D`W&jDk\7XD&?Y\9ȢG:${1`+i n8=%Ml%İȖb7AޗuV3A7ำqE*\qb'YpuHƩҬV nm=Ɂ-2=|5ʹ zi ' ׹U>8bK0%V\ t!Lku`+]c0h&)IVC)p| QUA:]XL/2La[Xѓ F;/-rtx-rei0hE˝ݸDt#{I} `v;jUvK S x1Q2XU&6k&lE"} Q\E)+u>.,SzbQ!g:l0r5aI`"Ǒm O\B!,ZDbjKM%q%Em(>Hm 2z=Eh^&hBk X%t>g:Y #)#vǷOV't d1 =_SEp+%L1OUaY쎹aZNnDZ6fV{r&ȑ|X!|i*FJT+gj׾,$'qg%HWc\4@'@—>9V*E :lw)e6;KK{s`>3X: P/%d1ؑHͦ4;W\hx锎vgqcU!}xF^jc5?7Ua,X nʬ^Cv'A$ƝKA`d;_/EZ~'*"ȜH*Duƽ˳bKg^raͭ̍*tPu*9bJ_ ;3It+v;3O'CX}k:U{⧘pvzz0V Y3'Dco\:^dnJF7a)AH v_§gbȩ<+S%EasUNfB7™:%GY \LXg3۾4\.?}f kj· dM[CaVۿ$XD'QǛU>UݸoRR?x^TE.1߬VwխmLaF݄",Uy%ífz,/o/Z^]ݖF\\UR7򱺹...m/~q[ /7n!7xB[)9nI [GۿsH\ow!>66}եl?|i [%۾s& Z&el-ɬeb.E)բA l1O,dE>-KjLOgeΏe|Bf".ax)֒t0E)J\8ʁ,Gulʂ+lh)6tqd!eó5d ¢ku|M"kP-&ђ5h ^pN0[|B>+q"/[ڲ&6!%<@fpѻKQ31pxFP>TU?!$VQ`Rc1wM "U8V15> =҆#xɮ}U`۸ہt=|X!~Pu(UeS@%Nb:.SZ1d!~\<}LY aBRJ@ѥuȑz.# 3tl7 ]وb Xnݔ[TN1|ttc‡-5=VrPhE0Ǐ}Wd|\aD;(;Ha.]1-{s1`HbKV$n}Z+sz'ʀ*E%N3o2c06JZW?V g>ed\)g.C]pj|4逜*@ nBID f"!!*7kS4޷V+8弔*A19`RI/Hй qPq3TY'퀜+/Ĥ'cp2\1: 0mtH,.7>\hSؗ΀ѩ آSNEYdEcaLF&"FhQ|![gIK v~,Jc%+8[dI368fp*CDrc3k.2WM:UbX[cO;R`RA]d+w!e rr솜[/V`+@;Τ`5d0ϕ_Lع`C"cK>JG.}Ε00e>& 2䯫vNj31c$ i '2Sn-51Y}rE~b>|Ď6Oj~ebIapul9| 3QtUqSCxTD7U9/nq.JYCtuc nrCtVDƖϧ;INOKx%'t+sFUJq:ǫf!NRT1D(3.8Q;І?O+JL0SU%jfˬ1lމZ|VA/.ȍȱh M-r ~[0AG꠭y*8D*-Rz_z{/S[*"꫒?`a;N6uilLn<Yllmb rY״͆jqTI!j.Pٱh s!:W_´KxA|Hk1nE6=W|$O -{]1Ak$ ѫQ6Plp;3F$RveL l5`:~@c>q,7}VE-Q8W70up˳ A¦g/OEU:غA>?=CۣPqȅlW11/$f*0@б 2Dݘrt +qrx!8 J&[V =͋A,z`S,J|L/vrʑ=}IhM4fG(Ȋ1{TT%41Oa'$ ]ms۸+|_*[dn7s4ws;svv2HٺDU/NiR:ںq{l @"5&IU1OD懻l)CU\4s@QHp6E+CU9wCUEWa8T(O*eQ&جJ;4B(H9\\\V1jVMYfc*a=K.?I!T$r S&B>g` D<>XK U)qGjH`5˂8TEq䏱h,\(RVr=69Oc;0yuU#Sh&ORuxOxxLOLǰE5=x>L!pyOr. 0;ǩ=T $S))SRbtOlS7HĄ9K2+0׷}|-̈8~Y\9Ox/+׶h,<䍐Y~81^WVD#nV9nq^i"&;u=[ŊQȲĮ9#*l=r~tWyߗ/tCA7)<ox"a(?=ɴ,2 ʱEdďH`yv Y1wӗ귃_Ju wzE2] 1uQ<].jRꦚO oIYW|Ko^QbY2x$W2~t֤Tש}KJO:XGfJHG&iXLALl7u% N՝O0:+n{BpS..yx\8_P0JGgOY#UͣZw?<)V,硼uw3 !?3}oPx|]9Ok۲}I G6NQĽ '2hDR,Rju*T_s=w>αLjSx='x${jg~czt̬U]V#H KZPyΨ>䧝~QSP({gWɴujQ"X, kwryo6XKLD7BBXkw(2v.y@!ΈްR%oXT}0 ADP8Щn1wүs'I>t'].ɝৗT'[z;C9˫d:zΛNNhvFƧKwޟ`8>Áwsyt˦V,c@ ŇPY ^%N ݸ|!9ѐ &}@ؽ]H N tK/݈o#k8jCx߂mwr&! r٠`xp+w[2SRMeE{Pæ( !ᐿ!N2zwa}4 ?-1aʥs&QYzl  ZM"oW+r=&;ҠDRvR+5%:+ۍEw"AG~˹u\eQLK4W}l:BE弇rLiPqڝY(GxKx۰Tl|v }HcZOsՄ1[W3M J qCʥEaQ%eJڸr/w,+RsAm攑AegoOqv`IuBE$m¦9i 56 N/{ huQT\=]ʥ5e ~'LYPSW`_eyg ̚ >8O`0x8dϒך:g! ֣M+~}0JGcihBFu Gq|ɦ2U#ePqXi_gd]Ĉ#M>>L׋ EQ"Iys|B6MТ̬pH6B"˾`:>:Io٢-,VZM?CpNNhR2UܨtRlupg (U PH֊'иmݶIp==UQ2FjooEp|#Xې -'Y25A&` UNӑ|5pc8Zاi}a "ĤeXŁe?V86:ZJݗ31>O*Ynn?noy[RS2v׮SsV*g`kHx{J3QE@Wos @e*XOW7pmGOYy9iOpmtEcDqe|RU쇶j#N$oc f.׵\ApdKՐ!eE:ښ4UZy5TJNaA34C}r"At+!$DG0e$b6ˣKl9jDEræ8tj  IdRb(,`Cq&g^ "Xk4$[S=-a_YKܖe; )K2$ږn2M)dyt3Sہg{q]R@L͚7Ķ@uS<Q麕6i.K `\`ڞN9b*=#i,IU-+ %[nQG(DP'JXӱ44(`ĥԈ\6\8:s3fčCs5W㙱ikYRE:b ˄OqOYY}#Lr~g^{g;[WK Gutbx:BUo YZwuW_?}55PƈVVyc$]kgd]{+Ș%әQ`d NK4\p*o7ڍ.=ߧqT] DM!&Wt"ہ{HH]ي`OӵV0|+ Ծa;Ppu Zj˦dH I,M{Ǽ:%zx*J! lA[͖XV &U+@Bl4'y6 \!N۲K,bY.>;3ǗbWt1&utnkop[?lwl!.bi+U/JiQwǯZTXx5}F}"HdH-КIC1bϠbݵDf9̌*R' p:+lAGhukhF! |"˜ӱi0UV,fvׄ8+lq<P UXi]t: ba.{r_ŭby9#T*̭; Q9cؘI!D7itfW ZFMRlx r ԂLۦLd?=&yݴ ANN7d5݂XU,Փue-]o=F,Lt12BHshk?#_ERҒ8:k2?E5/Yi-~ut<*|fgJkn,gw?2Qsoqt٦[fl:}^L7xOM)aYq~YGLO2{VS4#jiNDWZ~ўr[dL!b7'˴j^' wGM+؃ mԴݴnF=~.cInՐj0 _e&5vCMt6Z t?h @{2'Z~IGԢ fN3;x UP̔5bbF5պY-RGAAkv3 lr$Y }EmĄQq_@l]Ȫyd_R۷e4[Bm&s'M;1ⱩeN@AA  (c# 3w^.llI +Ć6.JU ܝIY,F3U7x//> `޵0mk鿂$/v\QONڝi2l6KJv=)aWݽMI$|'p-kWAԍEL"r,ڂ$<|~)JO'ůb޿yC.C,]yM9iHicŔc:k3xt/S(aKG#>{7z%2FL0_Ig)yYۗ28|Jȧ)B{L=Wٯߚ/dV~C 'f4./3ԽdH3!PWJU'JR)hk+YG-o 4IV7gTyɑg}~Rzq|㓟ru;0uLkvg"t; ٺ&;}֠y#Ʒ Evpȧp< "H7gwDQ13Z֡5tSb@U|@#ђy@h"Z4# I@D7* C6; 8v7\TB6Vcˁc: 8.ٱHGbA0@'ާrTN{v ̭pm"Fg6 ށUa`i7@0+~缭Ha&MgJD;L0ռ]7bbN@1\곈PD%R(2q: @fמc0ȹDvw +<{]Ȭ@wvFN\N>2@qgv ӑ|Lų91b]"0>FL!@ NCTސ_ c=QnBy,J_^_\7op^i_|7ߘwBZ70= ;";xNÎD,;0,!XxݳBnWuD cw˂ 1cv˿ppS>(,1͟ CcjSfbrlQ1yNWsbbyL^\Ҷ c9ڲ\S:6q!oۍȡe7n ?N0n E/?A!g^Wap:MʸAçF6ZvЅ8_HBmE31DMӢ=; (MV/7q}An~bmܹfNg.8Mޑ*n򚝧&oi74SM8qv# S/ߍ3N ,O©JYJع Qȩũ)qr~ї˯_KD `[Ѓjs)RRNLQjb庄Lj }<6#'k}@a*ˢHVG,.X$WDH4@۸bDl Hee쎈c;Avzppw+'`hI`ȥLƄ02ȼ $$ @(5+ru}fu7a+1)@އ&,1ݓYdTbm s^*?&V]wt5RL-S^Q㚚Z韛tDX] g^<%ܺGTu J1OSnB`TsAX2CR]%1C8dݑcWX3ұ7.UdGV=PA6OKZs4'PtЕiHu̮e}z`v kҼ<w5. @,ZY>';?#"I<sXJ+T lQ?Z7 ](<oT}"?sA-],WWn<&dF\ާС;Z#>͒yt Ǣ*K߃4a ;#(п'䖒RZ  27rm0Gyװ9:9c2ÊńǙIڹNd9zf4s q^CRU}dU<HC-͕?qF|px|6 8%N uzC7z̉y:L J4Hm{l(vEJ3I3 C:&|`-thImLHt F'?!N*%Iu'{qƉPl2ɧ8P٨"CQ~dAVk3U[ד}{۫VF-^]˗ ,0V x*sFU[kz\[ .UڏUP`CJ7}'sl=5a T@F~l-׎V e톎6l>_iZ/`Kv198 v.) 1Ak#vM) H&q8AqQ^#O/gQ}J=p<q(_>/>]Fuv路)Y .@maFφ bd Y`&#/ 2?Wڋm7yÄ3 (7#Y0_c4H ukS^!NXUOcvhF^,ě>ŬQfiz+b+ܢU#o@L 7`[ON @~nqDeX.xWe?LQ[r0M=ao{ЅII,c[~C(]%@(ݝPztOB. )eLJUҝu}`w}б{pw]gZgbiu#ٓRgBUBuw'=PwOBV v ۝P0B= v _%߁PwB$߁``B #4ؓ`B*lBe{v w N(?P'|BUB w'4B*$%# c0*,"q4#5.+}LWɡ"'~z}9}+W1D8f廪.c<'YW9.Q%(o;ԆM^4( is$=q~qdT=W{AuϮ|,-)"`(siud>͢?E,2/y2Atch=)"tkAS3F2DmfY"JSDCuj\:DqfFIl ^e6!,a  !Жch/d=Y\>D(b Fĝ c}Dj1Qo,TjퟳzO/x!H">j0n #O}hM'\ ~aӋ,D>Nj-r,}]f1;,bV@ouG V6!XHIWMn&@TeJ 跗Uãv,/@{. n [ 5Mgja>*<~'9p6]4qaLpWAzWm_Y}Χ\5{)Q\]ٜHFS##8kၱ6@] 0@s8hJ$4gH AU^ cVJ%.h:b͡D>B>#@ŜVxSo)a֨r1A}.&Br07LkV@AhLi]JAKd\֒+y䝼Wh<]{^ ~8e8|A[p8BKD\3"^#k1ͿD "5bU( O͌$j;JN~qJ~'kHrwMAXNSa -0\CDʐW6ix<Gc* h몫-a%KQ PښvW",$W2P@rf1Z;TaѤmXB#ԚRhjZCkg~ 4ab\5#g _ʒh>z"gbj˧?kαeB1ג>|+%ֱ;! =[N]^`/ pi6\\b)xI,9#(HƆGBmKND9&4ae6G, bR%P(6&{LL'aMlϽUrtG±hTѹi] <;F&nߡpj$v -"AqWi|SKM@ü%;5gD1u,}‰6v֍ n ^6]^;2՘څ!>#E1#Ȁ0=wp0r a'M{fH<8f#yI@4dJDGw"wG~}E i)c<~J Qmc_A6퉿d?Կ2Nj%zHg:<,ؗ=y_ O%)J5ER{d6Jb[qqzc?Ԃt*}܂ӫ7v?]5v7զ7n>_ 6<7qzz_&)'P7wvrфOO7Z> JKk KK;? jctzjz&wbW3[?O'=i6<}6=Gf\`7ڡj~5Z(9g7?gC'fzP~:Ě߅m#{{B[v[kڕv`; mWOo%&m% (<Ĭ=l,OHOIcejM>_."  Hxô"ʪGޜ%/P;Ϡ=lXYi-338H3IhJGaP2<7pUL(|PJ.m{U>Q$/u517G`N‹Tb2o G3;(/ $"FXTv)`0 =B@ xaI"YC$^rPX&%yxIs/skYhYAAiu>j FHz ;3٠b )985R8K]['$h3n!@Cѹj=!j7fK͍KKL81p_<>i9ZAq-$T4c2)F #h3ZYюtmiN/m /)M+ELW[F`+&2 90b/k֪&34H:77]p|1<]-"oAV@MX4a- !&W4ьH66NƩɠ{kUrk^ 1HI~ CX%sW TBtYr ".4)eG;G]|B3itPJDOs.i"Eޅ@~x+usF|c%\:C Ln6ح (?Vf,F3_<%|m ϧ Z"Mv)f5R.:( OJH\8gj>=|XwକYxJNiU؞WLBEN*mR+#I*BgVqU9pW%,:&iGsm tRe,52lusH:3©uaļu;#γQcgjJzPƷqmw  %N(=!Ql\L ƺA*0*hjie_BÈ Al遼o*vщExL=܏k=@lBW2UdQ3mՒLHjQ$XouT ,567Dr6qV4)b=d_ 7+Td3x 3 *^Si(bě涟 SL *rP+˄Eѐg5[Hg~ yN쵄StBjj8N7.8\x)Uf>Ƀ'\Ԏ"M)H?5V#Z`4P/rpl(v`:#>rkH9Gy~!&e0DT VVo%̓Q;S9t,ŲZNH*wS~RT2tλ$*-08C$dVuN$KiHrP3>GT)C3]HmDD=<ͬ[TtI5GKm4TYfCE:g Ě@k<H^K9/UwE sw\}KTýjG,K2s4H0ހ U@-dE.zU9$x\]/%lDFNOCt-_TLrz=Gz < mUv?݈{|70lb)%D|)}q1o }tzQI!ݏIU.nH=rQtIAjG,!onn˷. 4:R ر_0(euN'?ٷvw XٖlYL1gޢQ.OP:{JEU`o f٣kʫE!x_׻,2e}e>:#<`އPW'~Y]pcnSf m(Qpj&)H-_&ڌm,A^9N[2" zi^o^6.0M.KR`T3Xo0 NPEѭ*IB,Q$/7KzXle]>`֎{WT*մyvy=-f-{xN`A C#+BXq,ЃלּSͥyanOz?ֳ~zX?]c|UYkBuI]ON򨒯%/14+<q΋8"za}מFF4n*8-mJD\g֋r~/><~`jN,th_ʢV8צ06J+шq %>EL(cĵY2epvx//]ptR ND?]+u(J褩wIg8YuExwُҫE90Ň9F*ߣO«2qi}k6Go#jG-C[ "Ok.o]^_[ eG+U#<(6unT͍c-֪L[x -g5J^^؅1FC[ -bR"hS?^6I OuG-G6#@99֯ct ܅\cpt6vW`? 澿܆,v= e{q F}N=|`;ׯ7núЩlYm3Bikv0-1ֽU|QQ'#Cjzodb fY6y ࣪]Rvʮ׻-Zv0}e~3ʰ\zаy0\δ!2]'ǖA%o@E! `&i0q^k4Dt)WqK)(k* 4=&'%:.9g4e]H7KG/1Vr%MZ!ϔ ΅8Ftn_oo8zڅ]7mhƛ~p<E+P?_0= D oE_N͍CNH"E9 PU 2 n$\IÔ10+-2b8w D@syڕmAKGN\I#OzS|pҥ uܴtiȟnC:z(ũm^}nU]%|.Y8[ >F Cnت.3 sEL..1K9PZ8ߟe˻.8^qiCɂҍM"vc3s-T) k SA8_sL8CtNG^o8x^'I^q|V6.8jj,)g.8Z+_'j,$ Jm<=X#sV8'#L vR. nı"cś,FSPASq~,鼺"˺ B/jlVj ' {J|sV)F"ٻFn$4l.Ydl M`l/<._[)ˎn= {YUHz ̆@ }Za7;F&H G"Iʎ49T~-64aoT5 @ hMZm{ dTZɁ+o;"=+k*{BnNx|m}hqb_g匑QQӆ[Lطd &lr̦W;Py1!(T-`Kp&#Aܰ}VWƦITy;鞗 vZ >Z{袷!H|W޵=56"LK'a iSގ|XݵG/%${Ď- "=d 4x1;bV^t.&h (sJJTيL^;~82SJMMk_JAt`T>r:)?β,lMm+)";,/;/·gm܃k'hƒ|{bt0|7+" yjV݋"k.u# c3 c?xs]sY9n,Z*,~y\v1 d,dhL<(ӱOoX Řm]"N"hV.i . X*FJ DZ`Q3' z~3[%gCϿIzl<`K p)k^yz |3M̍ɭ~RU{3^ݞ$Kt>`OYe 5^jJT[\P/̄OV@'If+\n1z|So,O@t:)O ~Hz[ oǁ}96u>=}6d<+O$q0O$TjV+?'Gpgew*J=" Tѕ ;\.}g3Wv `^ќUՂD ?k;zhqTҢ;qt6hmb f+fC;31A ;cKRy[0g8 K5EE*w=+ו:Ŕ^O!p U>Pՙbˣ_0%~LyHDfxuiJ1k_y:FMj u IИƇBKSNaY=|x󰯩$ܙ UqཫMq5:X'd܊3Ki\Te \Q *Ϧz,DF,Jk9^O/-<~ǿ_9ˊXypqho%f~_%ʙf\!٣ Y9 ެi&=PZ!`Czr$Qf$%4 r`(%|MI!H%pg&:S(-p+1lZr=Y=Lmrx_GDo#,H/]ymq6P&a)FmH>[,w-0_ ze;J2 >PbcOBAX W,A=b BRγ_@ #m_$Gr֣q^!WT'h"hVR=Zq>*խ`| Zh7Oy9-M>aRPRWG4)+ )G˚ on~j-dz)Fezʪ O 6U/ UtaUF&yB/ȳ۷w}5ݽ CP97|-da#קAϣ_j j\N߄L`jD$fs/E\*N<nʤM=V,tށ4/L(3+R2YXʣ#|fͻxQX g#k-u"F'cYnvZz&|!F'ױ|Us#"Tmg _Aw™Q<,~C3qioʻ:xBϼ:xl.+EEt*Nm:{@~֪7~ .̼,iĢ "CE?əne~1.*U-MEv=yʏpuz}Мd2K98Ip6JZA3#$q2Ml뇶BռcyJ2׍e$0cGI|!ت z5C; ؛q> 'oP$3`_oG >`x[wYE:-',xCB_ه5̯)GZiWa|M)s53c%׍çC D;ws8g=LK3W@]N]tl|:O;*pD$љ}y@)KgnB&VS ݮBR*wS?nY?~Aقv4 vhʲ @GQЎd!%n]`~``$X(KDL݅ Lѡ%.Mn&D;R4 60/;KHx脽.9 (3N9򖏇k5#4|4("@.zn|-%KZdJ KKyM)Аv)i2̤2K&sS%`Dk5wF GDUG*$efᅢ)M$\;%$i|*4/ e0f/ Pkˇi26Vfx*dc @咦HJQGZJQmmei~_3 }9/k,sL10􊥆 PMfiSCqZGbۦogZ=>[F;*A%&=\&> ܕl6|bLF;Q-O KB 5m*bRU`@AZ%[k&6Ʊ=6plўADQP KT PG4䵵xYpuqHZj2H,Q ~1[i*WY9"Ĵ-dͶYre?ZA)zHWq) I"K eÙav aSa3Bg Ic4LZk#:#b&N<=?^fxEY2!{|k1e7mkM7?7Ndh)󵗦WRa6g D=( P &*`(:F#2s\b顶7}pH`|7Wa4ᔿojaF;yꍀɹ220Ш> Wz b)I v jo a94SX@@B8D()R6Mx PcX?A2ER * jE; ݷGJ&d@`c@\Eb4Bc)S׎s+Rp9Kī\+U4$A<HNW1+#M;QhX@%lP,R–) <05`5jj$a@ L](&ioTetICYe)(BMdP̘IRZD X1 Y=ƒ%DLp&ck+s(KpqK(AH1fcgJx8_!JTEdAcΘϺ)9zDvqH=a9dBR*3i24e&vG 쨀="(Z(#9ҦՉUO(̤~,i)YZP0U|qjʎ5XibMw-.uw's4Y]|+=Fc0Ъx*ՈTh$d `Ndl#*TSflb/`",[J/i|8}1 L)IA-}t eToH*d,#SaWXUIase0$7Y)DA3-)̖4YeFO)A0ա+ٙѕbKYc#/(d`i4Mt?JrV236X3XC+I q ;K7fswT(n'NM'Nez<^+VJMT]M{Y$PCьyEXxEIO<-źjب((әIPvGؽ|:0.Ye-"I90籍WD T$\n߲6|DVI4u=n(NppyHKH 3 4P]ݸ~9{N1z}x7*:{|(z"+W$+*կ>Jr|<1Jb,3j!snJgW7Nqf[|xx8wzVmRuUg3_)JF 7"z6+?}*ze 2ʓ}dn%dƮ8OHkv: '[xɵ TY*!7y%;^5J9J/ vJ'ۿׁn߁+|*f<ٳVw^MS Imm0?ubvYd>yϑ+ٿSGGrt>8'WJ[)y#9ގj¥ wFOSn{qn2ܓj~wxOms []Rf䪘o@<;Kn@gWӛARqӢ<ߒ&(/ti}1eRXǃ? /f>g2o@IoNP\ί~H[}\[v["ǜ)bOJƟ(y\kPC=48c=Tp৷/=mYr˯o|>xuc_;g[ӿ&-B~q ~zm/,bSo|zy.fgm2 }ك7SƕP?8~50;*kUrv9Sĭg#-yړًX8+Ìdyy1=p YrzP=yVc;&l ;*p=W../*A\.ݫsګ%}~Ô릾M ܮ5^'GK`{]j$s/rzq6z8V\& ެ:!]'CO`vSkǷ9,><0k:k=cR8V_{T]NγFa%C5&b|ss'Ry?Fy`ZhsTjZvy?p̽_ˑ/*pvo8\wgWE ӫË0o^|~}87<}+s8`X?~@*)w-.$ϋ'@3ivSl_ߜ]\d tt:XdϚV {EWR09Xc&:5^6)`2?zRWume|'vhcMuu9X4cv:΂%g:%X!ڻ[C۱5faq_^cfAW<2jݝ33L()꒵.5|Z\.XI IsV'}5nO1tn;:aj.Eea]k.` o堄qhbx1G0gwnYlfGd_;l0hs60p=^:G2)@Sөv˝v+';S. ^ "ŠUMZH`/Juc0]{iq(4"G~,Ϗ//N'ٍg.'d˥$}~lyYqw '?^beg>z;(Ȋ"_|*|U 5NGv+@䵖t[Q31LVr* E2"+YM[QKevs}^\^kv黜c-VZ:Bk"ń,f'&-,yJY²Lӂjk XYkJfd: ,=3gPpul@:ajGZq-',X\ ׇt/Їfа_D n wu`(h!; \';o  jN,J- L%!MdHjfvC>^A/`Ub ;l*:Kw)cm2ZdXdÑ):9Mf3$'S,nq~1` Ipl~2R gdDD|Y1wFP: heX}Q,Kh]KL{ 5a锇na%tx08qrSPDX<3UqW(R{)b ֡>Z1UAVh1!iH8>bE3DI(. !0gሺء;Z섻Xeٯ=3VqH$zlIX$\'Vۑ>6Q JOۛ>ce&}f3QKөӭFlgxE;2Ms:[8ܕ0"v_WڕJfTFF\PmJHVv%id\sjwd^Ffݚ8lX#% l.86q!an׏R"߃"yRqc $[AJ٥ 5l`;㈺=\Z SSQi!5i}4UZ &;?ط qp%]ma\]gմOt8 ()t̸5G0Y=77'yqIܳN;]>+3Qyz}CfIw'^xi%/T?U>J4@)Z~քYs|_7Ts!0u,scVQ g c~=cˬKˬe֢^f,l9qVħ SFp5]`O[~Y} y7f%G&PKdd#fTof$֦ZM ۛᇭ|CgW{@RVq:KD8Y&z~6# Rzׇu}8~wfSD Ii{&;񊆨3P)b]*f5uvڤD)_gRղ] 3e9q%r=s.~ D>D=@ [@r5/ұqҤ'0HH{g}R>F7Hi$Mi>pj<"?w0} ۙ>a't]GREyi/' mSsN1ݍ)YCgxcW7Vb1"q9QJ5.( '{ 4H5\nh&W`qNwd&#$-5 ZG߬F)ElVݰQh%g~L0)a΍~[{.ʺ9mS˼f nfWgP ^ T O)WIbJ&%,57JO*ŤKiC1\; k;x<\Siq[^T'xPu UvHr_Mj86tX6$x~ԃ NV|U26;L7I-GO)n)dgZGIQp3656{#`vk oaiUЎI<.*.ՙ%W9+rUDsJr)]«6OW,I X L!X@V<2X@Vr.CR+b36' )5""?tC`"C\+ jD:\z"HED:sm}RFݡ(4BaBh^TŜ}Czsk7כ[F˞r׺@^n*J /r+߈KF8^H)0*:#:r }ZA(*Ѧw}8'pWDU4AGDY|` y=P(&3 64Qu&ReoÓBټ`NtH!J}5:Q)-SÊ 4*Ev]uNf55<_ς|zIjj!s u%lPD) %:$`MuH6. :QC>ZS¾kFLE>L:@j2AC.ҧMRT(T ֩`9fPbuܑ1z rwry'P>qis̥eoWD(/ktU*tbBFZ_Բz&5d5wNPLBsG9|}Qh_yQF!>Z7yY8R-Xt#*]k [`5zL%ZO>q|7{I 6E;82uZ:^ђo[V?{GBdĜGű c5LR3N4jTFzD81pP +0sJxcb9}Ҽo5}ywץ󾿣yߔ9IBKv$1} S7O;c{.8>Gښoݙ򷣦kwtw?0mG,mT^p?{wxǫ_~lW~ѐt5-}VGOygn]u2'1]pPYMЃv|}{zL\*Eeir᫗./~~vo][WKYF̹Um̆IlgIW ]W,ۍD-s\UVzHv[Wet?49Kβj{, %S{>69 SV[m^)`t;E FV{דIrǸ*WD뭮]xJ-+H0WMo5Gcz=;{f.A_HPM^(nGK!TP]ouz)N*PLz=F+>]}H#cR "eWmoj{= @SĀ@r6ɸסa\O}j ] M Tϝ܊Tw=;쥍4z6{:GWyz`+I)A 8N(SuPx˹⺹qZmӝ/˫MV.ړzZOؚq+rF&Nc{LVXB hVsK3#}qH&T;*@^jBuj ӏ/ .aJhc tl6y\v_֎aF΍oۙ J-QjB[aPu;Ak6Qzfz#FT nGRì @,ͬu=FAf - !FRGzKU4M3F<6+$@n'@qр!i +^r/Բo CuDnb帨 TDWˋ[KR%ť ~ɜ× _Lf90|+{ƾ45!TH:,)vYڍimG%xclBD d)7f$c] 7<~a]Bva6Tyv. Fk.͘5{uk F]~QcRo';as'<aUvooD;!LҊ+Y R d =3?bbg.?N`b 2i f01te)\Ꙥ.MsiQ$F.M֋v%Ĩ!й,>(OUZȨ&]9Aʁk5;ՔP)޵ڭPK2':Ѹ8%oW'AJEpi uPDߴbkԞxIF!BNZP+(u~>|?]]EYD!;(P5T}ɮYd\ݏbm͂;qs]}Rw!z(f]3O)#H9QڡuhcopQd:`"-[18T1q(L4&$]Ӓz[NY*2_ R (9Yb՜%#gBoK{b"%%F"! lVq* g "G$D:9"RɓɊ m![O>r1Y_KzgWw7]ǯdz?tHIg?ꇳOH~ڞ덙ówvv~"G2~[=f oW9sx9$IH}kS$AiJO( ARAȝn5UGE@J^F U8p8 XAOXc?p[2E$BO2d^4HN G-LF71=a f7~}6e$de6H1`Ν,X'vލ`y ,MP\BbZ|:l!Gs>2"iuޕK$9 J|Rҙ-OKJXILCc#HOXЁ}C+K7nl0Xe^J F:JJNG"퀮?P Ry:ݖq:R:,&hb$ *Xs=A3p‚톮 ٻm$TCreKx?T罻&dSIjwRS u#KZQ3 %zXdfchth4g}5x `@WqIMP0~Xq7,a-TpeJuK] b>*4r9 t2KVʫ\ƕBB*BOF.ƕRdSI8{|a2͒u݈!Z}F[@n^݂O1|r`Ĵjk8t A~<#P`WqЎo~|hA^?}nuto<;bc&)FXQż= !#G5fyZh3Ρ*ZT4v-(Ƒ*[[91Pك L|۪@޶" ! alcCw& IJkZQؗ&E}a   lxaxJJ(԰g^)VYa܏0:i-.U#ШA~ Du;n5ic0Ujc ,{ٞGd@^_/`4svGz ("TX!k*ViE.yc8ԱK8|68\^Kh8^V÷{pJQZYe"NmkrVaRK7}:ؼ>y}蹼>0MKD`1,扨S^4|~I)/\iwfpcʜ c-̬NV+!P(05gTR ĥqAP8èѰKJ2Q^ܤoM{ 8-5qaexߣZ%ի~{GtOsv-3YjSw_JŢMœJKjEJ)Tn펱P=,V݈xz2\ٷ_e~:"iV.fꬱ/|sE&v9tK^s$VHR+_d> t\Q }q<K(" 9 }7/]L# u8/ҸH)րag{/ >~PJWH֦ h¥xp5g?>CO&ǮlrN:ǎV0|y!FC^I_&D,A&ŮmRXILK!%Vrƪ}pUx  p} .{ >ڞ/*ʟ= C{ 4Yrt{^_ǻкwxO `bP #701X+E83 Z&)CĔ?d8^<:/ϛ O3/4%.Ms%Oozه72_tqv3xFݥbja ! 5~aʋ ~y"$-:\5v4/ȏ /|z^{2 Q`$ޔBm>&ܴ ݭGUe?5 7mO`v7B!+Pi֩ja|07xD#^7h4=`e]hY; `|3΍nV[4z[Ҙ1ݖ'piN*hmh Wa-Cƈ141=.L ďK&k ,_[\yL2YSZ!J%v$AὛ(-Xj֑Xi40X s}JJ#bƊ!lHtL!RkɣSٽ0W,A[~.W9jyȦwwfV{>#xho)t8^ЊmFw7)FfyQ r‹nZ{3fMU߭x҆i {hv+v_up  si;9I_O;f_$•o»[1!ZH)cBXT3sseDŽ1FqoDG}eAZ"r SK`-L q9BdK 3<.q%YE1_"99@ׅ:f u:- ]tK,m<.NGNBH {Bcw[ W;{)@"j2z콖B#qb F>:%0-<'.0uerԯW%Ppn1 >$ `ȱZӺ==p[y56٭?Ҵ ߌaF e.kG8&o _m>s2TWv6E@jDXCyEqp:]\]۳Y,,+z$PCLG H6҆: jK9$_P•<#'g^\ ˘zZˢpY2&s96%:"Ɂ189RH#"={K_ iե7Cy/eEI`┭*Wa#">ahte:b2q#T=qőH9Ul.+J/5xh`ټ)J$o͑15E,i5+eYA )jnK{zS.C`kI)9:)0%5]'*9&}J+Rr͞*1bӨVMx ?@4;Rĩ#*W:Vĩ%G4).=naMi5=n%x΋BĨ٭|b6)ڞ;qul3mFjyG_qaYM=\1Ąe6塧‚ons@ݥX@إ}ZlBd=Dq I#NC5We[W5yt6B\#L~ϚPӭŽBEe XyN/x8)[CשJ a5[ck&(PR?9/0)D"pQ^XߒtoІ ߭xߊ0Puo`ݧʦjTKIk)RQbsl sMÖk) NW6&fX9%bP# '\8$@, .VnC'9F땽i`R*IRd [FX[do 1)1g&g@!d@ 1"V:g % Xa휅QM`04faݰZ46c!c"6J UTkr2Rm4<1Ba$$DƜLsqZ[_91`ƺ M+883 cJ`1S qBf KS:Mbc%lX4cfZצel%41AK੤0m6QF(X`8AUE) ^ũidT&Y\C)X!#f;pQ$q$$BOD` `cPRdž L'QW*x"e¤qB긹Q*8eXJ1rJ-6JV(@.86DI۴ 90ј]leF9 8u]O H IZ-Qj"'jR$ @MLd`!:a1bJ%&"HȲ‚?a'򬚥IBCV!#(E!i4JԤTZ`k Če,Xnu2>6Ci2L2 ئ1BssX^7Fdbג]` Ƃ0h ƓAe'[-J7!ct[5$]p*f8-RUZPb蘈($;ŗ,JhPJP](3WXktdaHcÐ.v@A`./m,{eF]IPn&(},ƒ1k6X*Zyput +RtP"ڌjF$zZ5^([o.HL@GFDd, CJ3 J)^ P?AjDlPQ,K'`Wf8#h3/H&H!(r"EPkʾS4K i-LE8Q5i@ ̂+›fT*b-Uy+XHK+z-w$a$R` Q}A_Ӓ%D0R ɘba2vyy[:,M ,YXbSX)>`h!7ۥgtt%)V F >Lʊ/ɟfWѭ6V'>hwnR.ssяW:_./gB4kkYV8Z~Gɢ\)q.Gw?{vS=V˾WtQ"4}Gr2J:@y5|FC%uQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQu Ru3v(Fe͟\j: Y:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըS:ըv:<QG@uQh5yF4uE4QuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuQuތQwKسgH~iZV{lWzoo!X-0= kJv<.W|v1|7\`h[[12X8S*Xr@rq(aǏ{E ,xh7V+ o쇻`(+؏k-7ů\ig4҅ӘSJrlG51ŒRhh`)dKsi:eJreA+P? +5Q Ҕh=ZҲF3EҲV3˳lGhnPR-˴X<|^Bl]%Ǎ#X~BZc@2j(`;̂uY[l(a -5|(aLi9Ƃ@Z˙>nb8rZaXf7f `UZ- :ed(l`6dt `-)ӣ;'0·&XF=6,' (l@J"J*,ow!Jo{``Fi>0UvX6x^!E] -uz5r{ eVrPs}KRJ ^Jx3ABXB#1u/ncr-΍!clEߧ40>/%߃9},;&otm+Y٢$Ix9O |xy>z?w.ͅ?|ozr R,ȍ槾{' _oAw,<+埓$#X1\m?]{w.qoT:˿SGT;U @RbɑQB XOw+ʾz `%ȫLv(Eز] JUk)g츏9` eɷ ]MK_~6_tgJp}j?V G.C'm9 oOmg>[{I3Vv] ]Z 7OWe,tgO\WСUAcUAihoSBKJCW.Cp*(;k*]};t%fOW, ] ZE:]]}t8SfHsWs` p4w;<,>6]K+]};ty9/q@tU]z0 h-!o JZꛤv7KA!evp^rqf%k;~ҎϟNףtՔw0+u:vڥOCY;߿-%o.[gq>Mg8,No^[E}JhY@>O}߾9t|,6E&i^<;_hi64R~ZGX&ՐhB m>ӣd&%T9(V4$>:IeMֆ!x=p}pZb1ErY&*o~wi -wW^ᓺ f1D6>נRsIs'WWW?,'t3>޳T ꣿAX/|n2{>^pۛǀ#B/^Mֿ:z8%9 ៷W> ؼw+;Xmp5m:FOt./B&Qߠ/mXmcb$-DsJ>PG3 6*b"AlTSs`:gn(]VA p1}-6Fէ(wx:^`=lz5<ڝ=O+̮&-|Mfe/i:m6_x =޸j?EyG<\^=;|ʰmRꖗ;W:vEѝ]~ϵ>]eqf28ҷ]Z n:O Sf\~9̹DO8| ^v4~Uρ_cʚ9_Z}؎w_ur( zgb@s٬VR03YWM2!TX@m ǵ?L[|ZH?*zя#,?"ˏ#,?"ˏȲRKʛن׻U',ү<.TX޶rn;Oe{OwӖ˄i v??K" >|?/BO}?Dn)Z:/wsŶKi{oz .;BNPGhrO)>ɖ z/|x=zN7|9/ݜ5O_?3p~x<=G3Ԧ7x&ZA8tAߜ*я7<B(9ѢTH,`U46ц?<Ω.yKﻙ^uiWkP3g#!e /~e?i+~<7YnoĠWl1'xq]˭.cwkݾܭ/WuYvTgY iU1M\(VN{&490iᗿMӧImd 鏭߿]גϪde\$[ +xp"z> 9Gr4 j6ȹttFdpZeIyeRғ1>#3|"=>5&g5s.8-d<^MQLp31O`ƑOm~ [;查B m!rжRŠXbԁ|W`|;7(ژAm ?D+|\EEC=PqSA5 Ȼ.u^Q8m{iI9SV3$"b3⹾" mCCg#8J8/k͂ .9a|l_D|5V7(|’/]9/=I9wDra {(zנ Yi9z1`F’5`rwx 8UQY _ܗLBE"f=8e~pdT5-vO73s h49^8~a{$XnXnm5Ɋ6(/}i}ytr{$?ݲmW9iV~2>+ٲQznۘIk;- *W/^= P<ub@tHuq;pZf5W#zP{@FˆZ&X{]2,H"uS#lp@^1Sϔ'Bnx;CLJIОi$N'JP.qiQg6 |*VFWw M tZ2@~f O ՅxI@-'~kGhd ~-I MRԉ 8@уG>rfᆏם^E 1+LШ|>i2~6E>'܌&Q1 DP3I]gh+%+%Qu7Pqh 9W,|hnT34ЅS:(cʎDØ!:&RC#uKq x:(+WQ7Dѥ$&jWi33#BQ' Er%$d*&5񒀣)A0DY],. ¸Ԑ)|i Y8ă}/E>\[0Lfb:1椄ɪDI<"b.q%E 'b++ʨ)L@mܙj nbpuz%9"\2&al>*rZ`NEp;+-U20 !|oF0}2Bf#& om7F%Zfs7gRwJQqG)𝒐3f^ZP1 %\Kb}?,dnW4 L.jQ7an0而Smٷۋ @WNwa"rS)-[ wn-E<'m68])$n=Qw[2&oQS!|^K= SQc0n0߆ +zBJL)Àkj]ح˘/ض;JuM"݊2?ŗ槴J3=$vn^*3"Gibs9 {MMU4_chG7ZVi{iIBTI#WR,4`y_^r>(.Z_Z\43O(d ؎汯c_*X+ 2N\c۶n5n*zƽ+@h`Q+_*ڻtkעF[?_/wN\:e4ܤR-yLX%IB{N u#y}d]y̳FI0O3S2sDyP1/uT8X3o\L[Eg力zv(WTl_xP7ǬW~%t=WMw$~D32j'(IA߀~" [7܎];fN0FI MwP"̂ғYFM<@<Կ@OUѺYX}Y٬x\DǔM[~1# tp O4ryhm̼4ѧ.cUGE>?n*Xb(8b,n zFjD5x p4TU%n$Ĭqɿb4a-! Jv[E%[j6 0"zP) υs~<ΡAKzEqf!1Sr"/|oS]}%:h*d,4th5bFIr}A߬}ɞ'mYNF2ءobS_Wb>+egi6J!Әħ=HqUҤ;㧜(#\F/(h'c(+ħIj9߄oM˱۝jv&6*cAj{oitn0kmW};1J%' !FlP8g8.hwǼ(E%w)(T_ұԽ1EOhAUT*5/C1F,D{fGu qCSHN~{x`f i*>8S6K R'Ԯ’ ݏd`*WǴq*فÚ='*srhUTܤLD]T'l/͎OGD AO\8jS7A Âks͆0|9衷8. p ɾ վICr73c jVGˏ`H*֌~f??z(;)9L[EUоHu,1FPLj!9S2[T]4VP}ɼNC`Y_nj,C-] E-_l0Ο}o6pe7Le{^ pA?iu^oi r%fЕc'k}0=Z5k7eX`m$ -‚77)'ư<]%RDE[| ?, 3m{U-Any[z%[6xlx3zV@)[~a_wnZT1,hr!{Ϳd`>LOfJ \췳2Ilm7QJdn1·uvxhqG9Lemov0[G?_뭃^7bNDe1mhZ@e9PpGos vhV4 X zve0m-¶?ӧ>bث[)~Jx-#A) 7yB]ʎ,ӤȲ<ՙ0Vm&(q4 1ڇ (cI.TFßK\i9)ٷ -_ J1zC.]xAoGPL%šdH3 |M>DZ$1VH?:G=cd6 aDj?儃Wido|)@.@Kr.Rj3$)5vʘgRģm?:ިbm}'.F4ZkS@5\-*:Ei8a& . %Ey:bUYڤEKڔGf 1Iߍꢠ*y,.3zh b[\ë Qc4R0" <)#+.(8jd\)`e_ ى̙uKm*rZZҪAf$|f l35DZ`X\#)N!#I % :5YzV6m0dmxy &%~sS MNř]{zRx(OAވ<1dL,,>%xRF624= ;Ȣ9 0;_ΰv"<ыs}{7MKP7Rz*{3L;=[?#)K^qBQjbijq؟-v% W;0긟}"lhx][p+ϓKBE2=1SKg( ψ#FKA4e v lc1&q pUxPyс8̻ pO;.C9X>7XZQy4qINJ]X!:O{l^Tת'&4M=-$G|Z16VԆARaMךIҁഔx_a`eAv . 7-}[@:?@D|x/zU,Zxן=e:=Bt6Ezwի@U]@*YU~VP dqzAE#oa}hփI—>AAfTʴ}^_pٵBVhPHc^-_N*_mv[ q@ؗE΅GhaZ\l\@b:58<颠Rܥ$TInLyTYY޹OXd噣z ?<\iVwk%]}5m-s+Q9,_ ÷Yb2Iw╧͂G䟀 ,>߄#wYQ+[x,v/SM,߇w m?nRDﻷ5[ƒva,7K%XΰpBX:]X4Ummx]}57]h/풁Nkݰ&rC:~[ ~xZ-]UJo/:#X0jkL9 lm_g\dlHM>g\`5>ffՕ,Bw8֛Wvn&Դ"aÑ/yz_S_0lDf82f; pם1ۧҥ1 U^j.BU} Ԝ};YA"zvT YHbR'!ph+[n؜q P][ +k7B$lS&䜅< Q],ܜK\l(\I,8>#XFliS !URfx_6K3l.^?텇{K'"Q{a6eh58-m^]ٌCfpwM(>Sk+Ti˱i..#z(.c ="_E0Vq8SR/PqR#F쉢|ip߼.|Hjٍe ۮY9 ܥLH"'ƒش3{ c&GOE@j9G1q2"2Fe[nJ<0%Ro3Rxjm uxf<_bQY$nbaYÁJ&cE17h0=B1,yK! @&)A'`4i*1c/} eè 3Fd8O5"OaS6۫O3uwj91&H=s#fg_hVwXtES1r÷qv,^9OD? 'EL2m,K;I< 24ѦDA$s)!7˦Owh%WCوװUr4Ρ$h,h0T4_WZKd"Ubj#=_x`:㱅k5w1x|[Xg+`}UF崄( =Tu2۔eԀ%\Nla BLZ/e]q b/Ѯ*fH)W+Yk`琛maїt8a6!{~S.!ֱ \Pts m0g1%$xsSC-q?xT8Msr4*"F9ui-*xXi}hA)efҀOwA9ZC/e`,Gi$tEuI~_Y+WxޮNr0ɅS+\X Pg(fBﲗg^ZʌN2x$I-W-8 h K-㋱%DSķ{VUKwN#֝<;YFHM bg`0#*PxyUFN Wx qYɅg .M9>Zp [,5*R TRhPXSb Ǥ v6(+a/K;g-WZpkM"6yBO5+ݼ~F1{z—(i`'Aov]Pm={WyaGhvV&d)IznK 4UɲPAaAomП^!-0X \O1>w:61DSdNJUx]cV  oi aq2#vpNQZpyi %)zyR:ÞX'NV5poY8rY8xD݉nT]vuބW^3 0:WYd>}{Vr~.cؗݞcOx<'[Ilqk6@&)Ӈ-)2 \^!d5Kȯ՘ _^\atɹPmhW=t~2jIM9+2--1lt.GF0&P'!o|>e3x5xj"ZƇ]~fSoǞ)1UT#];u>-m8]P/ax2n _-e7Aǔ]쨲zKR,*"#0mH)1OrpbMz{d|``cPR1qI.n%TMm94demV[dГ4*-k1fmn-3˹ُ>įϓ0vۗ?.2gqӢDݗ)+|x_EE?.ч?fv{q{6k\#َ'֥mmgXKG*e,XBYz4X~2 YgE}YƔe1a9,(֖ۢ$چ(OvR^C#|zB =v=ݶA1h7֛55S3uiu2 g3AFΖ=tq`?G܀_S-7jLcKYdB~qCOOF#( Y&\p\=cMi=]>v1>Bf/}*L5닆1 VfaLr3AZ+DB?HKp咉/-s%< <[`k0 | Q c^lB "\} QD1gom?z0|ЙKF R]t'+B$Y#iݨdQܟ>bR=vƯyM 30Ԣɨa0 fp;BM%s/-#EKi*\ y}ew2t(w.ɩYF{0<-RcE}^`tZ^ 6Jo5)Ƿdq(VxyoPT! {an/ ޸^JPB#8Dl%sxNb_X82YʻyTʗUsAHniBՌxdE¼+_23{u G*6{2jM P I^ܨ3!h[bw61,W{M -&ֱ$W4Ro j79;̙X{jm…eR|P^00a;q1Ajɨ!!)雃$䎽&AɨQ~fy VY8Ԉm v("kzA)L; ~(;IMɨr(1wxde2^Gow\PQj\D,Xȋ[b  3u.ڔ{gR[{H[C"Ԏ$rx+&) @9U5˵[ 0%GvBRv TwDxb~2*xKU21AVOF  *F{$†$k*KӢoq8Y:3n<χQcǓ>e>JVrWqrx2E8 ?*TdFFNˏi$߆~)1^m$0GxT9ХȻo69nl я_W+._WZOQ?0QX?hʩZ4l48ʡѼҽ=ӂ:d eɾLQeI4 +iP?b^y|JȞ*:O(zgśޏe*L^mԌĴCCQK婩䆾WPE{m!7p)LKZWpwp{A6NF4wDsC[LM.:l-Z'clXqVnrLw:+.R)QX\}Kf́Un[T[;gr9z[o*R{']t\}}$"qH,|GǨE`E#:c;0u8Y;nAo^%ۋPѐ\̵H k%gXb%󙀏- [|0>'BabiYls⎻c@=R1gY'/Kqfx]p@PYA|`C/OF )(QT:8Kaiњ5FpmK0G{uoAɨ-lg1fm(@K*RWRq>+fI=q'چ(OZbWVi{_ L*}T# ^]',-u ShJv&6Ɇ5 r)g,0ʨ,g_#6mDlPMٓQCSy/rZUHF7k<"ыHHch4ټ*|2I;_'|q>$mť"T PZrрIflZyăIf̓8(7{S0+*  %iϕo3 :QC)F= ɋ;U;$qug>C}N~C\.}K-]z a=|骸.589ޯN$켅*(ߢ//l%+8;V!&17۱H79]2pcL'g lC wz2;ZT8䧜. 5dۘo<&(a7Qe.yt"ܤLyΩ/ZrTNK'3NAƵ;X ArdT8dhU+= fkE3inrd) M:̻Z06G|Dzkt*%U$燅BFoq0rX\C޾<ҵל"0]%-dc@!oZC/%/6Xx I08:/zGxwŭe`9~ՕRخ!T:P8-4FhHt\]}x|{= !^z=SHC A:ܬ:k0Lc i<ý|6\KmOv٢. heQ;M-FJܙrPL4׆>$ѷ bl_>=/`{~2*KnQl@\K"jV5IE*@p ҷ ?D*?G:NH)Kӯ>O~ ޽8;^=mn({;_ on,2Dd *n`/ Y+OVl@EwMKh}vޥv 87K edRyY2G6}{dami@O===(`7 qty!%ΈF9  lT33Y!+le(58 l,S`Bqc1^)ؼ Y(6x`ʠp5(@/sl;kI2Zr{2WfYNǯ wPGL\6nR=tƛ魪0 ;u$bEb>6<&aQHYfkp3ZNԟ=nms\khTz9_lg9O؆W𜜟O= pfyШ /`<$a{W&COS'Uv$KAqAaڝ+A.L?8\ F" .ɫ {h.Kaۛ3"lL}keyLA2BU a(*h1MMɼ"-.={%WBWǷpx+CR@+9| @k T*c$pAT`Ձ+ u$[ERV-t]g'%ȧJ 5ycڰf ]dXkX5Q$2rLH knjeq`L /\D`|.(XC28D;7H!|n} uN!U~ڒFepMŒ!kI>daFepNd% [wi:msT\C28Iw. ]*68vg g{ӛbȬ4*#.4L0 +4L (5ߊ>^ K9ߧ0R6`Sȕ%1)6by?mI.8mx 'T^V-ކ,.7te4 bw'{ofh<|),1 ].&F?r*#R=T]=j麓=nP 61`caKl&S`LvAtÈ\Rxg ʤ:,)z0APMkemг:pѺuղNQ\㲯yVfBP:T .) -~ђFepHkgTiuC)'++beh6r?a˺XC28BM8Y1khT1zzzPu/b,r̺'d'5.: a۫sƽKwriޡCa MJ^Rc1|!~'ax$KBE2j~pg=DlVu9GFep UMeE}E (6 {֓ l/cV^?gYACJ Zā6ȺIRn J`ss0n`eSI2JTf[6S˛ROz9ίQUHkg]h{o5hԆs"w$+~ ;h=Ӷ=|l9Ш N cO 4 #Rj(y%苹AƑ6Ek(S#{YMyjK,ƽL*hUƋ7qʭRQjMJYmX(,wԒFepNbZg R~e]l Ek<p>la"1\JF|q__9;khTBz-yӶO )]Tk.fX2~r~!Ulof3[TH)Qb'o0*aKt}x 694P`Mʕl GFA*DuML!.pq]/$Z2 nR;Rvs|ИsHy쑘%HШ N< @wNexLwL]C28r즬}hfn(F]p0.~bRV" wr}*qQPSox\B\NeΒFep.h #U.-<{ۣyV|u!xFYAYkhTb~|QհQ(pjʗ,q(PBʙLfglΌ(ZNB 7$ś94 N4fݼry^>2.;9X xH,UMl$KJ9W(gEMjPz+9-O^u7Ͳ 낕^g vO;6̹18) R'+RƚWVS;-1n|胜khTi߼^^*[y#x*G/3ve0 T>7sm'.& #ЬIɅ!e(L8L#׎}h,ؕ{YC28|է3GҀBք.{BA1-iTG{eabr9pV&8Vh:P 54#LR@e,v];0R)NU ` d; $\i[e)qgǭ&,W|n&|4{1{'.۰tW>_l7[VyG8% \RCׅv136NTF68ؑ0v|":~lu/͜,h&&fL`yz3ADQ`uyV|Ik~/Lj>q-,HAbЎi7N }e]GO?i֣?y=<O?E}bcLwNжvn֤R]Iܓ/'Gv-@ۂ(mcW)/}-L}X̏#˩ @Lf75V`TxӽP;]mƓ,s2JϺ9/T-BQV>>yis~zo{;a ],k8%axG,4E:Fצso2䯃 }4Z .'pݾle ˶ w6J2irZ8StL3k<:EbRʏ`3# b9:跴7AMoї-ћ10o.ػI[uo 1 c5-C]sGрB#zj)1v"_As`-, S{ YOTC &Fn$zw8Vܱ-c[qp;Z^/7*ԏ_oM7 =|ڟ}Y?ϿuW?:Ls_I"^[8<4>\=IAW1j@"3keP0vY4a#A2rVk/;Sx->c}lXOxĚQFۗӨd>GᅬxO!X6 msN?3۴?b|HhK(ŏ||o#jdoȇFO.킱$eT4__>E5iY{6G@ŸX2=ݖ__ZlT7=i6ғ }GY7s׋w/}0"篦kbプzpѣ_2$͋t<凃ǯ?? 89gX+EtÒ&FuH)Ȇ gzóŦK r<"Hy냉KM-a$H8Hs?7f)C0D!e֠H[XA p!﵌FMFSYHKD,Mb篎N_|yp\`LQ-ҘUs>i2=,VCW. 4Z-@7뗿B}LƹB)@\`N `oW12#`939^82*s!՗ZK(IKB:fh/#R:"Eߜo?E::<:98[#g0_Oy3qT@Św?7oN{Ж I ߷X 85n$E& c$͸'G-`Uv؎e%udYe[Xa %yg=-߃d'Pf#@eX.Q> 6gjl0nㆰ|2|Q8g.X p~Da/m OEgLƿ{6o-yba.=^7q7v&#r俣ѧSM{ Ewv<y#AF?h~}u2>g@ka[z;|ak㵯ѿ0I^[IMW: r_\0aTn9܈-)ٍ/ z}R%m}dpu+)$S:C ;n阆A:?QMG/h Q[Gp }!#KGHwA'v.J45!%p|tzM95 ja% ypI7K?;_SKzt|/r;]']L, (s{Z;CGM??9qQ9\E>':uF!O+O4^?};:#xO1y$7 K)Gڕw ܉ ѕ7P3ACMtoeG8=[\ #9Xr3ng:oAZL1 ӵks~tS2/>7D+zg*"e5?}E禼rS&f{¿I~Ց dˇOaPV=M_2.4u@& λ9ac@Yp%›Ng9͖MΚ)*׷Vqy'{+K*+fYIYY.rz5tw<2Y xEL1@<C>x$19 \2$^v&\N Trovî]@y>QWv'hhk!!p[& ,aҥOcK_~1ɀupBkqY[vICt0].7[sddjnslӯG6WT&iN$NL, 1i>$Wle,J(-.]&3!߇RXf~V@/!ntNZB-0G̍ր vF=+ʨ&F@pcl(\ 6hSKN*PT}*r `.0@|! \.Ő| ۹< XP;ˆII _ȗ=!ȓ F~谾q)biXvCV8olO>Տr?;Am(G> qa8nKPNNYe ʿJ:ԡ Nn06OD0 pצ$`lM  L^悠PB=KQ)3~l̙J)F<>F~,Lugc^:>Vz,RKgJξ,= sU5P0Ǧhs:F Y hllp:-&}PCN*_8ua)VUe Eo."c$Ȫf*DVeIBbκ!1ѽ- Μ 2UTúPՖ/ekR6ƚ)xT),ހJT<6' 2ŊM1Fm>)Ia2 Dl(6'̎7aЈ@ķ}N"뚾+^T ,cm| w=C+;'|{^@H4:i7OFم%!JN[z #1n1Tit }4:FUJSitYcP[)LTB=*Nѩ4x KQf=ez@̆lvLr\^&L4T1tT1tU ]T1tU ]CW_:죮*YuC}*{Ye/+ԷTJV2IoU>5 lc**#ߔH-׽2RUQVT0TR#TU>BP#TU>;`-o8"lS G=57T7 3j G@_ʧ Eлu $(@thao+5RvqlW4cELJDx`z 4ۏ7%>Ա5rznsUb;{BF=Ai&@)gÃfi*tr+\c RZ;|UN4ǠlJpjfTZPz-.ıdawT"sZªي*EhJ*JUbciL"u ѢV'Kr9ɂ~B#(#rO?4R(P) "h7PvR "Mtg>k׻?3wZnZhoTC\2iFn[k%joBIl?vs\ZWLSjK {ffK ,!`um @D)6 0[6hp"h=DNO#K$p'UD$"L2b77>~ά&Uʩ"Y!8f lbJVL{-&ZN=Әe##4Pj=4LBw `J E14hh7BӴoj[O=Arg-\^/bq6VgYel Ou(7/i~4R!02_P>w"o2ѭ(?Ucw1cFG ~QDB` >"LDI0VDs˅GyϐSWkٛ:kTrK<'ՙ6H\6F\9Ʀ+]\!*BU+ץI@#\ך++2WF:à-Yp;/+A-l8XĻ |+ m! ui 3l#`cn.i;iטia'QaN@sOXKmR'!<߉iqoKPp/v->I9C3=&[6-1x3GQ}ۥ.w i="*H9"B}ۗhSjۊ"L3$&E}Z6-1x2Z 9q-&Y=L u q-+|%YM|a4`aK_:=hdZ>XnF9[=_f_m!edeijM0]\m˴PM rLCQ@(hli5h[Uʛ10"놰c=03E-eZc׭M[adϗi;en+ .اr]cc\H-]QWp]e3_#LoҺ ˰:_2Uz4D+oha|K2]V$nqSdVoK 3n: Pz\avK./ ׺0@~ȇASUy$0tFZCbJە 照+aFKL1Oɠ{^s~C 3GN;}AVHH2$iGN:$J$n𹶕8Χ5~g]4HԱSɁzgHTEt†:azY\ߴu_f,]]^\GV. >k/}T8}GZKH.Vm2F@j+[s[Z<>GHfW Kj O ºOAɅ5[⼸g7^$⒒iEI=\!P8Nѐ-0,h,+ [ZRƱhuSD}--pBN0S`6z#$֕9 Γ 4{?UYJEʶ@LNpR^q~C 3'Ϋ!;-o!w5Kr 췴0y+q6#a" @P=_*%i -pޅi`&d΋E΄DUmD[Z'3spds C67*̛EKK+'̷0yG> jjm-#Uh ޖf8/vY:ӑȫͲK _v90hxj/ RpHPr)}i:9S V-3};a,U  b.WϽq% 筙_k- ^-yX(UWDΖf8lU@q [Tڜ 罥ٰo`3! E@՞rS2練0yӵ}P#1GN2e--p>OS+DidXP+dJVΧgzՐL0KWI LI4q}ق--Lp> 0^%4=a-2ʧ~|X0O>`"{ /LW3ץ/KW]BWRJtء(p'DWӡaS+NWJLW"]D  o0WNF]dSat ҕKdO) ]\OTjXհ2gʏbv5  s$~OIWսdɧ{ ɾSR[+¯Y~[jv_}1^H\}XfX[rumfu}o)z0b[nVF)ȧ(50Gs!%cHub{PU+5fv:b[z|_Alţ{{]{n?+?ի__]9/ס%n~ig7g ٮF?f2n)9z$}P8'&KB]kj~} ?D@n3mwp v:87N>;o쭚 QbI"zsUM>| ֟~pIXLJHu [-d/pԧoQ񃽬i6camf[f)f aIlF[6|KW[nS d׶v|>poEz 8))UTrFnvag2,%dYre-MLb&G!{4̴w+3=Gg#ߓԖ}E$O͐Pv;m7J۷ciGpָ垏Y;G]šN 'zZ9COZaSEχCObB!{:GaƒeNW2#"].Sz;/'CW9Rkհ3]}to/vZf7g˿ n=Laog/Ir_ q;X3oWR߼,cϮ|>}7oޮ_+EѼa;xUg4J|Qߘǣׯ*lJ5g|n]hWzS~t^r~]58[B?߿t]vnwzqqYFE݉1/w۫_ܕ^,7WB .WعY6@?ߚr~;G7^m8[yoJoowwnj~L٭7g+vv;~GD_u|Q`TT1/Ʃw#<|!<5G댔+G # 8zȑq׻7$,돗+`=~׿᧜Z=uz6Sȉ3Qva!*Bx[2|kssg koOA]_t&JU`FV$!YKc#Jkj9Rjb\5I'+iϚ&dZ.%8%$1,}ٺQjo0PR '߈SœவmҚ 5xLf c6 } }r Eąi1UM*a#LXA\b^oW7#.Sڥl\t@n[4 tr*0 b̥a1{^Re 9b0,|Gj -;Jkl9o}m-fS H/FY%ƚ X !E`$ښX/Ie9[$[Lb>Q 75.Wlh>dkFi#:ڝT`:0f,(2AQc :`jϒnB,ڵR.!Lq;a5*i1cL9Xb ' v**5(:j`0O#4 ?v#zˣ4TD(QVέz<`ՇJ,dxE|vT:C8V1q)|XLĝdCqkyvzXWa+3\"ˮ ?*l@4F* G)WD節BAJ)`jK9f,)(3'lt_q p@*1K'6\`lp(H\=!D\ɔf  pR.CiZ|W2Tf0 #r,?lVLK.B@F7Wѻ*dWr0wBʈ͐jPoȻb+"~x7(c)Hx#(!4AHXP J@&T"M)tk eKykV@xGŜ ݄cbj " n,IPg# ,7֣97uD&:F*AC5썁D]`#(#5Y З_̻8YQ/1]˘`f̾ ]="U.I4'ry15AHIyr З;a{vw5Ù j(3mI!+jЎ@ȃL'^WHX|v;^[cS5W˽:@K3{VN(Pʍ!wdX#jGw `qE b W''.4 N:!khVbraBo((BTUR(LF|T51Q:}e~ Yav,9YASZhciznzo 5Xz |n*PqI<(bdPQWеkH|sP)Pxӵ@9$*\d! ^V0в5:Ƌ&@\b 4%7 0_C yZ5GA&Tu*% 滊P@ \4_j5fC(Fӥ#D`.VFNJYY jm?.a=ub} Pqu [1t^'QvEB@zjg,0B3;:H5?^.…ЯGj-n5Ϻ RK**MWy%HW "ګ%g!9=}fel7.J<ߢDS*BW.=/Pŵpb4|[C\1U'+L(uχ6_-q\8N' {? eٷMN Yq4Jq |G b6'&pl:VN)Riv{iԁlɨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:dy6t-+dd|P]7hu#F) uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Qgg:rcQr1\o1F@)D48+$uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Qgg:^8Q|:l:Fk4ɨR:mfԑ ȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:dFVC߻݊w2(ҧR=l o±mX.,-=hKOZ+b||Em+%_CT9&B^BWV t(5#CF]`Hò+kD.thuI=+\֕]`L6tpM6 ;OWRpGttd6'\l CW~<]Q0ٌ C* ]!\r+DܮӕƥJdDWX|6+\I]#]Y)ZdCW\ zt(DW{HWNB +ϛAx.th]!J½3+\!hm6tp}6S3T߰!C.~-PJ[tŷ+NtЦH'2+ +L.th57NWri%s3+,x>tpȅyBV]!]I͕ω s}2B:]!JG>yFt%•:B:]!JRW{IWZr { ]\!+Dj Ǯbp7OߌNqOLGo.bi J#^UB%*ۃOoF?!Ȏخ2 Ƨz.d<þ)5<5Ƹ*pQV^yVzncgZhvjOsV6?kD'܉ = gGI *WM"e+_6wTɕ7We0](jҾ|_0*xjz:j+2R߇}R{7mnP7%V: l7Z')~(4+?^ĥAwX=GЛqzi%?톖4-jaq%浢[2#(\g+]q퟽>uZE( :c+d0sFZ64q%P%("75yL =v.n6V^6XR+?FhdOgP-2&u]}h}O?TSF7ND*B:bDkŮtd}9t%6lzq,s;6U{/]mWg{+6xѕ؂Cۧ?Wl ]!zJOtt4ˆf"-AKP':m|v> s\ß hZ4iu[m~|4ŧJGM~ng: {0Za.̴mlZ "Zv\eDH͈cl J ]!ڎn;V]!])'b6:森STxvc]}Ni2+}6tpF]!ZewDW{HWY#sJmyBjzk$#磮=&5;OW>ҕ3܈~; #(P2Y gu ưnv>Z56,̟_o߾~~r>iNJ5ΕJ1]VɅ2: vwӿA?LЯ~@-b* m2[ypM#aHK w^~'>^>][фՃ*ⲗ,~g8ŏ_o̻ų޸V]r5ʛuUUџ$K?Z2]]UAۛ8.$\ pjNGxV^CVB[9= BīsGy7JV [ s3urբ<-[Y|MU?#\bqY4 6` ⯳p99IK[C8k9>ns4~~q ܒO 6}6oFF㏣G_1ܚ}q.iYa$lOqz gݗLpщ 0QiVG1Bc1^<Ί4-,-vPmMmwUn۴?n7FC-ny_…L*Ѱ[>x>.O_^xG P.7[<6c%S +NpݭoB6[qusIB/o~wH (퀿%+)k<`+LR}<߿+eTq}85-^oq\7:ΦGG\CJs|>)ި]PYq֢:Jq IjYAitoS%&'{[=K'Ճi[EZZqum#㺲ʭ]*;g/^~4]j{@{M:*!]U6:4ыD],Q}6)+z(R=qF2g|yɜYx96 W\( Z ״@a(`~l} ]2 `j8'=c)sTUe*Q~V&JyJDzYp3NM(I9Ԛ )u҇TF FqӨ4E^q* 荠9@zIWޫ۝[N>{.fc"+qa!Bi!j@)jM[Jth߫]jWcƕ*YVRp&,-y]F;!D/09>JV5ĺV655.QNxkkVi)"4Z:JC$uXu7Tߵr("'0Hms|1jEbX&oVŧ_[MŇ?t_ qK͗_VY$hu[Nղkiȵ+W\?>G'FZ~;_Or;C2jR:A[c3 'YZ tӀZldgh齳I3~vĽ5]}~Ǥk4S;GOCTvk`we-e-K FxRbW:vpu6 Z#Jgh1ًYL&7lzyȼgGWƭڟ+vhkv(ݎ}[Е$zh ƥ&B^BW7j|<]!JItt%tFdDW?U칏pNWJ=+Ō2&#;߶-+6{+yno(p d:[ a,{0K6)d-R>sΙጐ j @"B%Uwp%+h8+˃]Zeڎ+TilJQFdHv) W(WB>*%qE\i* )ԮYW W`+P+mq*{ 5ʀpY@3 `bW U>I\ nJb4z礨h&iY "I:Z.-tD~o{u7E3WC ŋ~I6MšuLQ]ARSK5ߊg=zfv~HWTSUYw*T}y +?,#<ƣZxCҖw==QŐ0Z4'Xs?5򸸪V#ŋ멤-=zS" \/ͮ% P$Wc0PxWM `pj)%mdUqa46 \qΔ c/m W+lI8 U,\Zcڎ+5);܊B(XjG~OǞ8WҪWĕ2@0:\\R~fT*Fz\uWxMP+,$ W\ Pn}Pi{\uWJ>Yũ/\+\y4ͧWlvMgE,>⼠d=fk~:}ЊN3Q,-VdoFԊ I8^+ʥ"ZRs]䪕V\`% g q*y?'|p%+V)&GU-#㪞\uݥꩵGZqQK%-Õ+jߪRXW+ P6 2Bێ+TImI+XHB: JBBeBB ZpM@ UXWֶߺB ;,ʕ:\Zڎ+P)I.J*EHvl54\ZAڎ+T)Uq6?68I@$'ź.iH!>5d&9+(ȕ^/?:?^'%5"gt bg11i7$ΜD|~ ySʳ(l/b,^{L);?Pp+,Xj:QDž_UEtr@uZ?SgKFj:7WjgxW fVƩ |nA-ؘ޸լڜ0tTbxH 7Ϛ07Z$)VKM[)-oRۚ&+m1;+{);h=~ ƈ#J;3vjqF2ka9=N&|gz3TIG20# >E>gI`qA`yR]QwؼODx)h> # cf2T:pCSiHo;ZOcTkY>JEx 4if+{m멵_V*9qE\1?&ff1gKВx|52S?R#\3+Բ2}-7R{f+MVT\553JG-Ry4J{R֒+ ţZ~7qh:XmlP{RZm g7+ Z%ێ+TJUz5wuϩ%X}zrϩViz*v+jߪFh6D+˂!5=:+ P ǺB>OZEiq*qA\qK* 2 Ph3*U vWr!T@B+ki( U޺"U;~l` p%&+PY+P)i"4Bʀp` c]ZEUq*y vWBrA0$\\+B ,_,†K3rm0 @bBz6^ [Z?uUOWS+ SZ+]WվUO+,U v\Jz\uWju@&  PmUqq8 V^4+U(ж UR㪃?+x p+kH(D $7ڏ+L="\ZP;l2=WBBrQp@ ,\Zm[Ip9$PX Jv\J㪋2\3XK04`pri03DB"7tz)y^-?¥^J>Z Rz?}ԪY Z\^Wr@1U},D/8_NRGjyŨhGc*Vj ֓+<_O>Zz*M˖f2=z*2 \`M=Ԃv\J*z\uWL S& 8z*qA\qɈ ?zxA\\CղBUq%WBQiH0~s @@ڎ+Pu vWRݲP\[Om4P;B=:+,oayp++ W޺BUq5A9 Xq P ևAUqeI>2`&y_WB;Yދ Xc4,s1$y)%MIb)*3!`fTj]Z-AYòv'fণ]j|4YϣEj],;o ϲH9Pꡛ|y#_2K~Y|768<%iBL1\&ڲ'w3NnV-?M#'V( +T%r'rYV> T=BOҊh) @,7 ՊֻIRa @ h%|ȸ W𡶶wVz;V?AqUGeظ'WٺXS[f]=zj92 \`T0B†+T{sjd `"nI^}o٪qٹ;B^quSىϤC %Lojm%W^+I% J੏='if)iׂPKVY1NTaE%\~Ƅ1*T i,QOi E,1g4@n<ӧn)u ANƫ|,r5CwMM2w f?v 6*x?E|]$>ߔf`\%`oRJfA~vzzH9.I+pt n>͓Z j?!jq`37YjS9xc.#11§ gw -#zY_?q&LƜWf=[@[}4_6g+d"ђ}iW&wn>X}ŃRu>?M?xtNz[y  P[[ҸvS|_Oqh}=7_V{'MKF`$ɺ>r=H.pSsOyTcr5O\xo }_=f;o#/Sٳ f_.à!|eٶ{[xHd9te%2;q1:+چ£[n@aa73N,zb1(C3?F%%sL(5Z$~,6`g30}C ( QQnlD9ѺttZђ}>8lRe>lrHDf|ʪ+ݸ)ҫIGW,ygX6n)un>R.fzjs^ - 8Z)CTS!6>eae<5$h;ݾ]Xw?Mu*>󿏠Y˻(wd48KxkŠh'D  π*Ewe||'/9zsgvqHon#F[l ;k萭A^׷x˟C4 v[uۡ ~v.>ʞbw[b\ټ%vÍ,zF8=$cHp!Q1=sR;Mɾ31v9- wLmOn?lWIʠ?{q /=By/3{ RثDT>ˋD|%Jzm $H\wgN*H(EQPml)Qb"S$H^f;"vyQ-zR7yqK=p"-yaUE IƃW.|ta: {x'=Л@ɛrN)8!4ID~\M7d9L3ҚC>,*sz_IJNIFs-SX:'MEr,fSK1[Ja{ li68 }M٠L~<pKZ};om2();; vvl׵wShg ^=VaNӇ;[~2S|*}y b (#@:Y?{ﳖ{3?=^r/PܽWxeg >L #&Ei!Z&6\$QMxngLLbke(E*%&Z|\p\{+zY> T|MDz~Y{Mu&OP('iؘTD!" kJt/ rC]͕MNQuMd:IHkRZXD 0b ,{CpǹBAqˤ,ZJ\ceYt%YHS^yd"[|8䬭Q>owKnxNCאSiV;[f~6.xM~5ڎGFis#r4΋e;dsmü!t .[C!?}`--; sx? nyw{6H:$ %Bck:鴜t'xzT~|oQAjQ$Do"$+pk =Ee2[f]Kg!<I0[#YJ noN߬4,f6>!W߿mj޲zjߗ'! Z\h niηx_N/]'g[x^IOFkEP/1'ar-q|GBtĢs$#v suel &&㫕7cNG&tJ=j]v͵]]wIӶ^V6f_.^NmXGrpؙkQ8ˏOij7\J'a:h;7¨ē<ȫ2.NmGAZ~6 TS5,8, |O?r_XYIq^MO'  o}ġ1jhhM^ |q-9q1nDN܅8p}맦N;s:S^UϷo:dR1gZYZ,u 1]F2HVڀIݗYyؔ)=%ʎMuIj.qIKjh6O2RaWClHdP/;Po:tẓd1v$j <v i2tE"k?W$6}"uүHJt"^H"^hn_?w;ݸ^Vlg"B)L},#*s-8aTF*&R[ŶޣܟWd4 yJkגѐmI %[} Ygz@6`H=G-f#>+j=ywW@]w3{w/Yaיƥii7hZK~oP=֝NgE?b0vosJ/o*Ptvv~MaZhS)Q$Y@'F#;v-vm`c:[&]+cUk o75&I.Wp)ULBxYMT\Ɋkӆ,`M$3X 5y.GU +ZdeHZ6Bbb*e*Qɒل5d1IbR& kZ(%pP(mL%'9:т֖J*'\tҗJ:Hido5Pc:xE٭h!ZPENgP.]9娴RV KL\$"D6 PtlƌeJ mǬltLHR$5ٚ ܣLå^.f:*c2[(Q !cs7p\iФ4IehPA+ Fw=q.UVgks6_ WjSh!}{.F Otڧ2;TnOY8B*aGhy1q1OhlmRl"1~w A6JҊ+C Zf ?' BI 2cEmY8e J,?}R\!MRBw ESps .~I.޴zl=I#uq1锴#mp2AD')`!hxh5"aB+@] ^%+ !D!: h{(a} (l]+$2]0@똌'x>-Ob׬|Z#`C*ѥq8V䅶@d˃&7Qb0'hI ![J@ 3h,a;&L͠j`oȻ\I1#~00c)HsgeP ^ b ?%Q`)IA1ΘecD U VҀ'ӒA ro.!M8# Fw3GPF`WɒpZ 4"|!oAAO)"!Gve"Bu.Zyʺ@J"V_MhDkDӰX\jU!PHȗ"cDh"2)V} ZDGDG-oaXy劵&!kE~,\uXgL҄n"<5Xum%k)F·JI1E%ŊԨIJ$B] r.` tg7`Nz+ta=XzmW28)ZPL.h8[VASc2Rg b )t^CooTA"&$Eȫ 䢢R0-ndFSOm_Tɠթ 7VDnh ^ ,:+*`Jɩ] YR~j[MyYI%Н M<}~sgl9TD`b ,k[4)Qw`҆{]Yzs 7D6y ܗU,U\׀R"d$e`wzw}r -hG;fCd]2t< is{+kFW(5# Na9RDڦ xd*.`mTW'2ID(&ʀARQp; d%C^isQVuEB"*BӉ䳦j9OwmY~ZLK~d &ɗz5MjIJ3SM,Q/l 2}SVթM:Bk͢:KiԀ,xV#UuzˬGae,BKIρ$Dxt^BecubHi~&kҘG!9}[ߗUޔTj+ojӨj2{CwkEO<~E7|d2FuoGGb\5E֚UQƮ[xТU0g)g* I df8J Uf0J 5/\4ut@Ǩ)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@ǫrR[;$%p(W(ֱW:B%g\:@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H tJ ?pCR0wQ0l(J @Ǩ(:@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H tJ )Y I VG F $W:F%{"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ Rzt_\Zoů7?ORZny}^\^PʝJɋպQ]NpI+.0.R/\JHt ¥Ϛ^h+]|0tn8':5R Otut% g+\BWK骢J͐γ`Uv(tUڃQtutM]8 p fӝC+]%]#]3e *`? p9CVUEc]]Y+ҖJf3v &{stRStut嬶jH q9p ]U࣫]#]U]I+\BW=uWx? st`k~p+]ֽRtJR AWMϝvJ ؿj9h=t(&:B9DW+|0tUqtUQ:Atut%=r+> `.`|(tȳ*JɉTݜ *0]UV2vtUQ EtutSj 'CWniBW@k>tJk=ҕeUwz.]Y&NtU -骢d(1c1˼ ]z(t<]c %])iݍ;].Ǔu>Zi ^ꖎS{9'ՖQ[ysyNg8-N?K%k(_*gST_L=L/Kmqпf%(_}+zhn|qwGkZQV}gyiE,7j$DףVf/sLV=3d_poLqַRForԣCP>L=gfPpf=܈d4@+9(a2Ȧ'+__`[bWW=bDWOmzn9wC+Z *\+BWׇNW@)&:B'm/ r(t%E}tUQjFtutUϫϥ+v$rC֊C6'*tRB XK1p`Ʈ*:}j?iW+-NuO?}fZ-͛p1.sk3O |6*&]-7>ڣew~jM?/lA7ѷ!p\\NZ+^o;k,/t2ҤpgAB fmcIpʇq6ջ!04zXa2]tu \?dաnOugp؜wE"o^NOr?ϫ/bojsu`ٮ͵|NOT=Od YQ:eqXHXoodns.yΛVҤ--&pO~b/ ?VS
^QJ^a 9]+t96( -^ sP?q{&xb ǭoxv%:It,|z ,67>iחlϝ6-f{z |i^Hor}jnؕ$uB8h9&ˬT>ݶZ-wVyg+d^ptce^ťێ&^N&OeK,ny;e4nL+4'^iqDݸ!BY !c];z!uHG[yV)q󤆎;`ͭ ݺapۤ[o]m۱#zx ^)sI9 g3>v? v4uf=s[|4~~ڙo᥍j}j"+IA*yfYRY|&D#$\pH_)흣‹[^}U.R$}(ZGZ[MZN%wΥVhUl=/3eN.k`Dfi܋n ͫVwݲ݋s'}[iˣcg ]w ʁ_'6n{P wէsJJYz_xh Qyk]bg"/遥Z/h3|8 aYLkQ_m&fS;'-;"iՆfI$JxIm.% s6 H#Ce+W6gRMѸ"eBXQFXʕ|9Ϋz0V`VLп۲3ګ@ǩA[Oj 0"/W>~/?ƫqKyjyc֜ a"C)VE_D:pD=zEIAGɆ<7mi`SrRk(QLdX.y:s]pf֫i3X&Z%pZϤII)ZW˞78ί~;$7{Pgo2T\0r t%*:!RK7%Ŋ)1-u ^ir ٢}D~eGbTlM5Lfmu%{9"lP^ Uj}qd@(-J I'm'VJ:yZy1 )SjGI M(WɅs`,%O q,♴s_&neH9h|~14UA'֣[fE"mc6%[-b,6ZxRXf#"v9/MU|SO|z|p\7\hs2Ej~+FKъɅ$L{cƧVLm$=6Xf;zz;1^.jǚ4n)Yn=4-uyXlM~&ye'ҝN@;^@M^i|k澵d4Ehv"E"P(J5J1ݠ7MD!/hqkmTd}VI  9S1;|aI֙5\D.HwmItiK4kGس#{aG (Ե1} @D"[aS"3ޫ|5se 2q;ĥsv.^U'TQiT{>%+LD~T ]sMsUT]\qK߸Ŝj4sBIVCd؃岹{ zg]WkY +Z\'\)9adkr+07ǔkܝb'2H:~\Z0ڂѵr{-(|Jm\i i). ;78~o$Z+QV5g~h?8*$T2bEΧwtاobL.Nב(!%bͭf}U|{KbyW~֥a籢{4-htD&K"Or46m^\Xd%pJ̊hXђAJƾ{ATr,Ҏ?ky-{d2 2I0Be/Ç@Ë_&u<8JASϜ%Z拡6Hr[d MRxokIݮwG3ޟ1/=l!ocCB& (Q:mScWh2:B'D`=jꠄ'"]tιZ|[)A$SAO [͂8FٖF(xsg}69 W[F+|q?yG\k| ǯ3(GT3rs‡?>/ߒW>p]SW-dzy r7_s>~MU`HA:g&8!1P@CF'"|d4pVdt15^rp`Ntr +0po_W/0@* 9|{iC{-^yp?= _|t*jɮ|L6h8ǹ5 64$7&R[jJ|Rh->K퓇&XCyY"$ 5g1҈ Z+kroʛHWNyȑ.mT/9#Ѹ!0XCr.prm")sy=$=Ո$/WB$G#qR"./ODHf臁)t1R8t?#8e4e[U26rQ _"S\(I6%K!<ƓTAn8mDS ƙ%r8\Tc3K Q)C%cTp5fΚqvxCz3;|(R]R)jå&(c5Ļ$Dc'h ٠pam4O&5=?1ylo&|uTRIPk] (ge&g=SjHwJ/jwqxXW%:Fl8:s}RN4t+ mϾ2'ԍ~" ǥP"v_*HewY׭?c鋫Aq%t._ht$E@o438^gRFYlv=٠#ZԼb{j1 Jm|FIoP ]>gGpHJc}s$hrYct^ڤC^帕0t&;JmvkQx7a#{_pvkw9I-+!&@Џ%ngfC?5f' ZoMmhB).2|H ʊJc.(]Զt"*NY[/Ǫ\61%FZS 1PKwQ!;c$:'-uڵAa5[7'h}X\qbU ٟFt}Kf=Yf4">/ߑўjݟfN мyX e-*U˭ͭ]—攐8v]t%h7wTfk2ty]~-uG`X2sW5eHe)xZL+i[?W F[06]vO@ȲY?u-sb"=!-{@ZsC9\u3A ]Xd%pJ^kfE4 hA]c_p{(Uª?kqԭw—Ki T|MHOI{~YA2B0kn~.7.7ou$f9ΰai8A ݞh!ݳ,ּK\c%IIOMƢ/k'tK%u;Vw܉+-~ݝ(=G_8 \ V\wq\= h:\vNb6$@Y(Mh tB[0IL+ܘ}[&\]܁߁Qdzf&5" TxѐPq6)BFБB Hgf1\h&6F pga}!2aeeh д6FSܯ1$&ntcD{䚄H.]n&04AoҞt(wq1A,BlBÌP&)I\$82Of?,$m7(Ub Dd#@dFAx ldgsDϬ3{2B$\ʎyfgad_*f>J04.|sRޓQyrҙ{o{B0_Ɏ` %SD.ѹbd}4ΫwJ`HVtb !BCw'k`HϝU'ΒaqÌ;?e'ِ:4!}Y4x:u/R6H?IM HQ98u\H@f'!5<"]KMzp:VtFsf/a/JR}۹Dմvza3%CG_MiD4̼Ea8*lp0q1yѰ]3zrC7M ܥ*Aփl4V(IZGahRW_1Kc|Q_ s3S4K՗WC}oFȉT-X\Q|+h4,zUM+FȚ !./@IgO:o}s3Lwg߼4^ԅGuGuGt`Nߵ'==Iդb-Hbx mr tXa7v_a@)/<[mw^{mM̋-|cPCE+5}~22ϚƾA;;G`;7 frnqzAc# %RIn;SPa8Jp*\:;R['XkvIӮ5s+.baʃ?m6bm'8j- \E9`["ԓ1x&Bu7/ :ZVo/:[٤kRW>?n`Uрm(_Rbr@ۮ!D!8bs juhoPgi*"R A!4nDy+tW=|C5"P6t;[1)[( +m)q&0C+壪Oe,:u~? :<y,![nu '1beP].\s3C:c&&VQJ!zƌƁZ1&+j$;Ƽ+?%R|zO;9#h})>Jб}~Sb`nZ R 23+*+u{ʘ5xL8J [fEg4ޥ%s>7WӋ*lug V8LzjXm[XB^jY:Vl&[l@tF!*VSUB͌̆`O;leքn)Jט|ܪvi/9vW(򹖜N)bxЄ?{6Pi:{ѡ5 C!fH Y@SJx-`0U;jȣf{Atk6}F?^snA;絧s^Z"MwuŻ]]xWW+ށ5#a*{yV%a n4i§fS~Ee?7f*+z]׏0g[O#?\+[=Ff^=BpU,`g՝~zqRV8  $&OȄw46 mtK&m _؆݌$|Tge'՛}MmVfwW4'@ ;mV.__+%YBܰG(i̮GWYWӅTܬB&>qT-w[xAȶ滾Wztz&*qNի[n X`T9F$㑅`"RSFDD b FрG!eLD醦rG7_ںDr_{*C9oժ"*jonG ^x wh5Z`(3*h2:ڀmT 5! _ic\ЫV;X~E=VFƼQ/Gk7?w9(Fj AYmhZmb. uEL@_͉rfsץ)ҋlI}Pv‚<ݓO7,Z 3͓Ř2(:o;cƌL"^ˈiDhnDl,\q03M~:3 I EX,@^vGE`.0 "- T)Jۄ7_lbWjC;-3;-VD\!&wwr0sAVZ~(6K%Z |nGm ^Ȍ ^`A ƜZ]{.ΡӮ5Q[yrֵ͖LtLgF=LxdN0+D{-"#a:0- 96at^iXBg `*8S:$( r|-(!<0GdRӱNǞi@ QzHLdu4(&E1g6c IN:M4mvmR]BCCX⯮?nyI}zZٯ?)J1/ ׁHs˕a:*Ț[V'=`gVؿ-C9']d+k̏=([}taKRs^>/AgF[d1w\ Xt\B%J+9"003L%GR$\.%6!`BLJ;f2Rʃp-[/5^q鞻r}&+ Ͼ]CW(o;szs9V|yKӆݸDY`y5i;7L˭8YuNT *A{BSM픣L3(B Q1L3qP:bPu$sr׿ZVE HitL*"Q* J3KЈ`@W =ƴU])뜲=X1JL"X4, ΄gL$8NR՝E'CI\SEz D":)SbH9vX9!PYp ބ#y"mV.޺pɯz1>ldgsDϬ3{2B$\yf!qK}5c()ҸIL??/S."A|ѽKݣeݘQ>d dMҽ+Rr 0EKA)L&a`G)QzW dE-6`MH"tH]ɬsgEvx~( 382>d ٥) I%Or*5hqәGW@nд+i||v֘ft/7ŋWWӋ<8[͉uf`;(yK oE jwWΘ× &J5RJ7{4h"CHHf^Xރ"06 8R|}#[2-r{cr }{$>t)ZuQm,Wnƫ xƫA X+0`j?\n⃜+[urϧ(E%VbR9c6([8S`{(nHT}fw (r0͆ήXI} U8p2ycnLS{/kMj.JZvgGSh_ ޡl[x-n Jo;U86x2jbEJzZB7KKԌ¼Dڪ'i{A7idLwȓq n&#W$R1[s^9mIrԇCI7I7cJ-lH%9qb3F?:viLKL>rhڷA.zIoJ~h@jϹyפLJJ-Ѣ 6;˜hJ)+Қ 􁗺Wa0L{0mCn{rFӾn"^W8Cv jf1cHaXA2V0*5m9k, +a$Y.-icsaNֳ>2XKP9|՝tduO %AJ*]Q%u?;vFoa9k DFNA>CB% ZlJŌ,FTZʒssVUdztVwr]'@DnGre,4uWsF&S-YHgxf9 88c4 I(C.M/0AiJ+PR24].V~GAY?13KHq"gi }Lw_ߺHa\\O%ٸc7^2j0 cLW}85H< %aL})IpGuQOzяɷf2 d 5Rp'XAz dw~]?JySpa1(e]-nyFΪşUi]BtEE |Pa466cxmY\^nmL^ˣHֻ /q?U~QuEOU沃D-u='|g~9:(^]Q^+qa)M;č ؟ 0gVj 6?c яg'զ,l4ӟ;(MtF@t(qdQҢJg(sQKbK>H:Eo-*״_Ė^+ r 8M(rneZGEdQ0A[R.k%!RHD#5h͖5f=6dS_=*;-g9g"z{pjh 7溹x\kչQO?ܠ+l lP>ڇjL{2UE#QHYu2LwZ D佖1EchnE*ˆldfmhD&˫2H8h[[]GXQK:6jw4żj5!պ 7vƦS8Nkub4l;+nV{mnwˍ7o !ӛWcG`4{5[rhxqQߐ~-2ۛݸa䞽:.ES'?ΕWpnl^13co'fҙq†;1LĢ08|.&pnHbh~_\.*PR|~+UB~fdz,_K >O -F ndF ndЄY4IVhb1: ?= )E5LP1)6[u-1ůi4oSiR8W㫼oR0)L`k &̫>+N+k.)oygSh hM{J@<|LQ$"K,h$\Z{[ 1?󻋹T]+~nYe쾟L.B.Z#wX'}t/إkBwO^'@(&Ћ3uiAɕe.%\*#PʐN>'kM̰}:9Z'@g! i*Ws@*'pt~ՠ2*~ݬ Wd\d%ew'OaF2RǠYJz4tx1O`U-Q.52r(#åf R3\jKp.5åf c~9 pѮaMNJ}~j1lepw8o赗2]Yf-YDfX:B .oymeq<zdiFyH XKvtL{L|Ba:e MMңt}cƾ*.yXO_ō lO-L ;VqŖ(m 0Q3 [&aDVKM)$eGҎY | \KLR5 kٳn% S퓳iY$]{+d@T'Ҥe@B%aNe"1tW L!AF306lKב H)XRY]٦03,*͸c  (B' L)< n@qqB0NFkʺ NJi)9 $DiY Ϙ Ip2 u!:eiil4$#e\NhJ>*ŔsVrp#Un€ARLݱ¹}9a{].64-9r5iʔ!bmLf_po//jg[ќ18gjR}FDm ֿ`+C'Z{b|{Omݐnh),oA}O5*MdoG)h꙳D\X)aRp@."FݡOGIx 8N01!h!P!m[nd MSQd P]BTah)L1WQg_Dcek oTW{soBhWu !f,X " -e}n~;ĉ( 8i88Bc 0*IPב:k֔z.hCL$j?ygLDeBHw8sޛ=GIPt }Z}כɅ,/(% #=Ҫ 4P&w7]JgW$Y(ךY +ziBpOe=U:c) ݜjQ=:\0O,۴7.qRH\w?YFLk^%pֱ$IO@*/Hb&}9%mEIvLQ߽sbB;dz #E79'?;IhD_@cqu%m"E_ _ {rGXzx$( }Y믑\qy]$i4Cs=4;D -`!ԉMIԦ]IR?|~UkO`Y0ǃ@$ Ž &dHfEy:>.״n2|WwtD6NMVgىׇϮo"M "~{CErY#TcFxI.Rg:EKbLx-2AvAke̙zڂΖ?rxUFّ2Ms |<8%wj'R}Y˹?~z6w) 'ْJ* ԏnʣ=>_GdW?9<@ G<]@B'! NQDk{ӺhVI,0ӷeݥ[](|g폦pRۭYmwklˮT_ڻveI+~4*mhv=)|3x;=0Zn'>LgkXjCb:oXlwg5PٌLrjeny'!w2KY3OZ?}a3W7;8RXQqyGvd^yY G0p/lwDjjM2AΖn*v*:8 Pؓz_)j g][5+mHv`RA nhuIRq`h3b&[&_ʺJKgGS†ޔF ΌZ1K1`tcdV5@D׻^[=zԣSoQ+/dȻ).mϛ"}Mu|Xe++ws^\u,/Fy{MRb1J&ijb (w[ %QMnpRp)߸GsE!:I&qAqV40@a(H(-GpHҊq*N)}tܪΚ|Mʬ.-*E8Ѷ2k>ok~r165%S>.FCjBH-,r +Bd 6X[J؄\ }*ʡk><}Np~Hҳv{ב]77nY/T"'fh#9Y \*/IEMb`"|L+݅ X1{9kOÒzL) <*d,2%ژKBxDQN"!<'cFSBQ&IEyPRf$eTG8@q(x99lL1³ߧQ[8<8 ~'_,mڧŻ.cB+4:VhD/ϡ607 kJ9]Bg4HV 62j,2Ҡ̉LLwZ*H\@D hf Jq&Τ&rT@ 㷞u\lGR('XjDʻd9|f'8ͅcj56Ɠo [[a:X ED NH-} ĐDщ + B+e{ykvJ?w=2bBh-)zapK(v|7Goe17&Ʋomڼ-pɣ_\CGoR gD>hbNE:Tܧues9x(A #TRj"ȝ "tJ Rpc)EK>"o $YKk lG: 3/ޟJqzzǂzƨ Nu*˼xO1@Ǔw ˃]I ^F߿>yG|tp)x;<o='y]<|fryDϋLPݹe= :⻄(~Bifr@*;@W K9'7Z3^eeR'/9PNfhBﺓS_g'אONe!V-5ͧJrr>0 盞oNo|| ðs+%#͵7566S{1ɝv0/%9s g3sŲ9.X=S)zxBˊw^]J Tjװyp]Ӄh.d\M, I#^ҭۦF {P)Y*!c~}ÈdzUxso Mj[3Rq oY4Vޒ+`N[XEDWPz:ڥ⁡҂΃cb4VtY(.Ѕ]y74#N.[Tg L,&2Gv Nh &iiߟ_~E#r̭#ǷGfѳ,Puo8 ess찇j[ƑRDa /4*2N:%@:RIqTvHЈ1RHXNଶq. Pɻ M:Npv6dAg<8d Tǖ\]dV!=%i6Y[e9e=*yFp#&EW 5i$ceƑ(2#L|[;$OsZO|r )R'J3^%)XHڀF+JrDQr9@ܞ<$XUեV7_}̺ r'F7H!HR *PV*I:)>|ORM;]2AK9W.$A{!S@3Me)jA-#:.%1bBKbkт^n`翣,R29Ёބꨅ)3Zsx/\`[ 8(I_ Y6azg}U>[yȥyS=%P'M[7 l)HU•}nD^]n.\!]/N_aMf \ͧ6ޣՍ&Pi~,ͩzP}q?7ߕ/],51`n( qmF?ʬToh7qiI֞@鶭ލ,OȈ!ƥ̫5xt,pir^I6Wxi\ ϟJ|"~A_}X0l?ܿzN}o|x$ig(Q0̪(( B+\ʵ2<+ 0UjQzw$_7}|.K?黯Qve}NF6?5ݻ^kvZv:zwWASnO+qm(EF q@\|ג%q=86FiM.*fp4PhcF_o@y t/.ƾHgmED<1",2^u<#o)}FR3g`B~i5r ]J9-Z MRD,u>i$aQxBrEhICET-7sjl&㩃(tAɈK.ڜ>|[ևtFK#dceǂ'h# $>,^EUu%7V+Vk0`ou+3{!(ЮB\RX\!lHZD.j[8 4NYmH"yHQ,q"4p4q*,a`U 'EZS 1PKw1AxG'u6R*; !y<xmwGC%{h/I* [/nVrёБf0W#\b$9@NPH\wm#YݘЮ@0MLg`A=Ӛ([FERl-[U?{MؚNIQN)93ZF1OzL0&dj4['_Z:P_^ F٣+v6\w'i iQV%Dv}0Udk vh)~Ylϲo~~/ҊWv_H~hS &K;pz~2u٪ 4G HOJh VRI)7IUŭOY>OZbvϴH 5Q?_ek$Ht,a_ w%O¯[=鰣r7[8'[oZ ^ ړo&ڡ_Fvh=w[|@|`%f:)+C[,;P\OZl=AV:Bѽj`I+3dۅG07`=/?Ug zVM0C.,DDh>e׋Y>i(ۂT5S?=7.<\bޡ\M{S -is$Iǝf PKam;U]qF#uT`LFߦߖ$I 0|:>>QB gſʚgGn>Lg_0ǽh q= WW/"veq(JXY12(,ًCԤ)/2.Krg|Ppjdi kW$i$E-\ْ#fjm\ Vlf|n4YvTjvU\,gyp d/^ Y4{osQ H,@<k̾LgPfӏT9JN|b5ñU$ݜ &cո G%<@rFϵ,wJÃ&iuP?q0@aUPʼnp͉4>1΢t/:}DJypvvE0|tH^? 4r6_;@d sZ *gKr_+eﶝ.ZINRqZ3jL'`H &e Â!(F&f652W*ol. Y_=ޘz'.Ϭq9$ 7XF,B0#Bpo}0z)#"b1h#2&"}8JC豭X.fUwx݅l7If)#ssg!ϝAgӘG"h@QAsglrD`p''.L|c9qCMT"v#QDH>$TVR&f+갵l,E_U63DrS+~.@Koz ocp܂ R}˚emzYveMՀ<c9˹A{ Q82Nw:/X:A[¤!GzהˬXhRlw[.eV!V,>U;?e,e2S4byGm =L] 1'Hɥ1\ vԇL}g8* f"-LaGYBW0#L9 5ΊۗӅ`˃E+>oM޲{JCmk!-@LG eZ30{-#c!Fs+%?ZAdzb\N5/xʐP5jxT~8!,$S,,VJ\v8f})vn6j]/AD:dumGKin:_6\?֝ywKzB})n!Y|+8\ R&oW12#XP1g&vkOerw#J϶N;.6X,~4U/6 !7x4_ 19hdi'3G JR\B:`^e.D"J80̲O"88fSa2 A3 ('`'{i:yqvgeǞi@ QzHLdu4(&9b-O==9)6Uk!jgVuK{H^wqwN\od!yaG .fF3kT!4;BV0m>*\h4z\>Ff׾aPք\`bns45AR鄡vqۿRa}<+"VLd@&oٍ)`f/&KfЭg[^zizin6 VՒ6c^*sx YlC8v&Fw[HP25^6:?ĝYnv E:6)UC'"w6mggyV.e]]r{?pS >8kKj,Wim-o]z .:u._Bk ?Bcml6cn'F񊭷F{·?W4M8oyZp }ldn+6a&Rփud]gp ^6Mt"!]C\oDk!u y3nb2]-餀Ouk40k|$_0D d/`٧zjVGsPg88i$]A ɨ;ߚڬq.6SJ=,} OM&N!L~+qJ[o+iZmmʆӲ*O6U_SƲ Q e B$h^ΙemCZlrŌF.WZJЪFJ dW`'a|qqکK^/gK3oCW:#TMD1F7ˋ Mgxf9 Gg,!e}"U""7OI+PR2<]̣*P j4!aJ` ~\Ʉ# ߅@ܹk֦يm&f˔ܛ$oQ7K+{策׾ )&ʲ!]<$_H6fxdpa놖yA\k;DQt]Z`@fNA y#vOσ?YC\)5K$JjIDs3f=VB5AAPI(A/!`[D$2 RjDs$rŐ* oEFӎ¯Lso6CKm\@Hu$00Rk4r6M/)-r C`4yk .Х ||*_qQ`4 'Qze{-&L9rQe t>mZA-yҐ'{|9ZTwH=vHʘ|#SX8RF% 3L G'cWC=M(6v۹}r$P 6Gpiy"K0-@,LB0,Z)v a-,|Ⱦ;v>^9_ ` )Emtt[HR&g a "ǣܣ;XQ3 ;D!EyX7w-yqKd+~iױ*٢~]lQkܐ0WBb@&~2(l Bǹgt Dhs- ,w<2 5(zAAƚGY,>5{eۀKL!&Eȍ48pp # )</DK-]жGnh :Rmm{ap8o.]>#=N\I. U{r7eu+7b bNъ#}÷Ioy(LI̿7ʪxUB gſ-gH4!?ٗ,p~=~& ÆvyѮy2eF @Fzlp[Nß^ ם=|Pl}g)\խ7_ \k)Z,ϖűڪD,TISFs]fˏo?/ʛвWeX A8s{ե&&8/Y=Ԃҹ%ß_=j;^t #ikk]9<ʕ'lxX} *RKwϾ/Ly{5h7j[6x_mo:^abJKAr;OwۛqN;ٽul4\h6iyuƫEkwnt.'W[UEU;RI^"fz-bk\>ֶ|77 7y%E'yBKUgφ,*"%'`VĤa.`FKc_0L\F C]^*ԞzqfPʘ]#30Ot4݇a\a:稅k!6uW(d/B,]]_Z~ٲ'`ϯ^>oG{f:7>4HٵTr+ήxzk;˱va{6V<{SJt2M>-0ZJch(2'ѰgսɱKqFvpuQJGbCWJFO)\+q2 =pFKcWi + W!NI+Ʉ4\N%\h)R=\Q=1\i +V_i2ʿd߼wv}qI#QE6I &HǜDjr͍$pEXd<سl3Ɩa}4'65Ÿ#wmI ewGeXd:$o'X @SbL IVoIq(R=MK3鞚:ꮃE# `<"2Zr0o@J#4x{73"s\ƂaU0ژ74=%bc^nF/^4Y W| _ʎ,ޙDHiNotO99aNH}H4vZ{O}rT1uUr'Zru qWK'!~?pYd6EÏhY]`9ۄ5xꌉ:&|ZOXéu<ƀm4C]];$I!iP1 eh)浠Jax*aH1(D*4(5;7͈%t{jA ȒiKĕ:VNzlBHRd>>1Hjٔn'#oRSauҒw"WՈ)6\Il%/*uY E)h < b*2YHK$zKCAy$JhTW(‰i+Xsߛ =_@*pzn6ۣxWJB0}r{ " % a Aтs"2HUE4!U.6O6bD$s#,xL֢%8 D{^-L ρ1DTR{*fI2#A ޗqe).Zb %-@ۜxH@؅h҆RWF\^12j=ĪYQ,~\>rW?~7?z5Mj\rvՋԃ˅S8<˰f|pHՋf,=Wh*8ͻX3=g$qnn={#) 7[4r[,CeYui c--ilyYE~'j}//6{/xWK(̓Ϫo0B+c9Ͳ-s` %p~ &4z.D~2L9؅Jy:鉶6Xyu:˨vg婨y>ݎ[2hzz{sx|´e=,Ϭ]75$GeğVnZ9J[r |+ 9e!0A)e]2L-3'6LjӰഁl6LO.zx <1jjcrKK୦{Z+1&[psʏ0C«=Oab>ixuStODl͜\)MMh.~t 9$f8fIlnN[l;B9c':iS|68b|jP>5lw' -46vz=NX;iа!>z?lZ]#E)e5tk&Q=+=g9-|_ tcbj!NAT>P;<>ӏDg;IWf8SU6F)D;MM.x뜡2DDy &g F8I!a>n]]7Z+QtTP$HςBS(t"8 jl:he]\Cn]*Yd!SS4o遅P;ޫۺhpwԩ ,}'G~jo6MujY`8eeÇO%Qr* -R8e^ q,tJ`UN/TEI V͎nsjk\Џz/(m Gt<0}1β^Ry NBPmc"^&t~BT;-( FcJAQ0HFD[B !GLMH؃VZdhJT5 $( Xʂ*g(nUUC2VJZ+iSNZ y+Wu ԇ`c,ƛUl{ma-U[L? J/Zkc@Y(MUC,A(L42Scu{;su2BNqQl#[5QHOڀe&5" K^hi4$/Td:uJt㠩ltHL(Em1siYm\D"FwAGhxl#1i<*.i..ϤK4Havc2㸶~Wvb~Ǜ@PZݢrdn$֤hJ&9@bq1G :(t0  "$x{;{K u5{&.gJFɃLJSV*hQ"Kw*;m+C!V=#?_a݄C LP F)QD $)@Whp@+$ZDkZ?z2M3rVK9W.$A{!S[k !F~ 5+HO lk.1!6mn,n~܊ gqg1C wR SW:_^SD$_d Y._axc}˼3,;LPB5oL_ڪqKd^bpF5:yq0J('솜]*9uΆ#TgkɹZh9+yDgcct`;$H_T ˋE7ӧT:η}b(lыw񮸱"P )j~&z|t}s>ʠ4(|\?ot9G3Jf6ߢ;僋{(Uq_g\]/ˌk\̛ėɁld`z6kK՘vnYf+ n$AjGzaX0J3,oQC+T̖M =s{8kGedӇY➓y!Ne$,}|!TszFxbMXi궜Z|w^39 n[{]B'ĽPW Qp=aӔUgyT;wnPHӷo/~՛//Oo.e?^_}%{E[iMSq a6zz847k`hSO |q4! ICB4bun6qd!N&2fpTziVӝ)}mk?k+":<da]շ?p(M=sh !䗆P#(@*إ3ܢ,Ф,Kĩ(rSܫ#SMXV1y1!h!!o[ndMSQd P mBU9H_c<񤌿3.[QЪIx-f;ുWKV`60V$[M_/(v]7r a+@*պ tQIQp"7+B6n-VA5l"2Q:,q"4p4q*,c`U kZur)\ІH;4 Θ ~fnqjᛯv4yYټGqrQL3ˣ.zTUna:3zr^|1cx оh\VIET}UQorr*g:EMBqt; %Pb\vNX$T{^4cx P14Ǒ$A  @ q곀U <5dCpT)m!Ձ #aQfaiH8(W (O%Ӛ`+u$ ?āiԡ|r '9ą9[znl $ &I@eZ^xF&G@w %p) weH&>>`0/ <]RIIvubSRJMY@b2/Ȉ/2->&EV *puDfz}?UPӦybnmSji{\fo*YΕ$Йdep0995 J8RX$ט3d2Ckg+=mp|> ɛSvLrd0T2@XX=Gam3tX;)-3FyBl|ǧNR][%KF@L"&4@)BbTxx\ԥ^eF,\9^$D>oT^"KiW"p(=Nt X/yA=+HΆrk1"4rL*aSdۘKrMArYN)8*4[&ɾ? nzvIVҲu*-d-LU))) C\$VIn! 2c ƺ0z*QMB^vqvV kpK=5&ݲ{ D?Y$`#֢>u.Rlԓ a'T :0y NU`;֙zU-hHQdz;˗:5yxɌkRwȤ 3 2PjI-]$Vj?_=BuKJf0xmJ#>刉Cf1ZKw;+^j i>d|_:7'+l*7sWssᥩf9IW{rf)ݍ+`h{ץOf# \p0 uuh{lG6c1f!e6|xw: {vy()n}d~2#X|uG|$ =y-gk\=g(]պo>:}[q 9:\eڜM2\{6uhSUFM c0l6GjO<:EY6(xd&FI@ L0bT&@Y\9a`[vGk7 $m¶Y6 fa,l=m+l%¼6rFN 9a#'lin#'lz 9a#'l䄍6rFNFN 9a#'lM9a#'l䄍6rFN؂M 6̠n-6rFN 9a#'l bVYq\Av%k]pjf*M+TLJ:D;Ckr'883Ti12|6Y_Qh:h (0F 5.t7_{5ex>x1_?+y2"I> E@"*2#y\Q[2X@#7sj̋O(i[./%`\?/țm:t}Ir)` a,}5 j޵^ilxijn:WGVoCM'dži@^5N1,Z6_\х?JֶWtzhsîr=o+x8݂3.Ɣ[فTH ?xK{֕.+^-Ul\6e"qLɉ p 4-~C`}o$lȗɌDIFx 1@$ Y6P:%HJ[emT[FktrZ9K!|AYz0'uk6?5?dmKH82*o`ʨ",3c 鐀G!|tR$ä5孢KvJk^,X6yz1Ǡ?.Rmo GcߐF&l>ŝ;#=b(mw? tKsKo |@:Af>AzIf5!%ZZV0EDt'mĘHs&LE4hWRXk٧sC[ǏvuЛu)O|L뙢nG_}(dst/bNB6!~NjߦI>SWx7;4|!{SŐ=R;\?֟'B ۤ7H+vL}: \]'; Qفa;ԓJ~0՘:>§4~j̅ͬBLꬋ+1n]ǂcu@/q0GcM;O~ ڨ!0dyЫf_PChW[cې֑b i->Y&u- HpRLcy vP~Of}i*飿ϓW Xs{2!+-%<+׍VÎv^7J %I/)&hd.d6pָLJKpB>(vωL,ICMc Bt׆jw\?]ϽaXtʯsE^,|_v:-ߟ/s۩+OtQs/Ӫ/H/Nn롖yZKWo5w "d)7fkЌF- L.C` 8,G|>B8L $SXAؼ^9kn7qܾS<#y@ }Bq2aunDcEacD;Wy @CX_g^$R0)*&r.X!M2\:Ƚr, ~+<]AG瓧B.te{$6s$c:G{ɸqVr:KCAg|t6Xƽr5WO6"Dm`E6 XL>/U#j&uVpG2EMɗr "ă8p]TF,$-jrPaoG}sbDՇzo)LI J*u&c$Nrt`Vlt*SPU2{Urh. t&Yq6}6LNk **kp8t2硂kg_w*ͳ]\׹phܨ\q:eZVp(58pU&AwLi(oes(w\Ò0A^x|1e!˅r:Mn|}~QO晱u!m޺2gsE#2an9$ȅyo9bYVg.r.4JU?kpʠ i@F3 l >hET~)AF`m@cB% u犔"ې;姘z B c+/ךPU6*X1bWݴܼv۫ j-99w|ۼ,a׷Xdrm`; =hp }{\΋9B \d11 *`:f!Mt 6Qq.ё&P¿S޴qUMellqߓZURvuG'Ѡsz6ok"pʈ4]&e]J^a`΄홱&*YЦ}o@HòTo$GdL9"iR&(1$ ɓv}!t;1k8cJ#$Qp"j\G &"jt l17Mk4퐚vvQ.rQ_vzUgA}Q"b{ 8RHU`EIǿv ZX@F9&[Kfl]<\t w{ѫOH[+c8,3Bz,؈zQ$\i5 `ZT MGhz˔BZs`!ZJ؀g2@ZMGmyl# R&/O6x UFr:24 懡ݸv "̥ky\wꖌ}*F0&Z$ fcGIJ'v$?u Y9מa**Rڀ.H02Xgm_Y&<+@~Hݴ)uK9W.Qч{F,Q#?V$5o$1!6Nm[^,nn5s!BxG2,_p^Zb8_I1Z50)"jU9䛲0ϭ̸kJ{J(B[i_e8bəqn )(9*9cwxJ߉=S^]%7ZOHt6TNm ~"u_eqn+ErE%oWϫ4ћ7G%WAspot=ҌFWɵ)Uqƚ.0#Z~^P]xw19p8[LϑoAuOϦ暶 "dI[  @ԯ(Y fYPC +T;pлz| NZܴJ u:uk՚kNS8 X 21Tb8=XVf~5NXΏO ݱi*&C׺}nފ2BfR rw\JViW c9wPH~GNwN('?}xG|/s:8l[!r nv-A#haвU_| \r _Ծ*цx!P~OOn#=We-S0HCC2<Όѷw/:M𓺃W CY[iO$+l {]')}NR3g`jHev) 18 4)KyT֩{L4<-drCۖ956Y&㩃(tAɈh P B|W*UL,ؾO/Y+/~QԎ&^--:'`|ɷ?(vվn|)餉;\ Jc.(]Զth=]Sy<ɚ#z'7BZ4 M1KXj$ĂF(lcG謷n~A};zծĥg5FdK|on:ɯ[虬# =Vdo&V}5t'VXvw8QUKBQW]E]el2Zu+`j"uѕ{3>caUҴ+8gWWCgVWD<♜{bcΠz諧J=RWfoUFW}QW4%2;TWLPl",uѥj_UƖ]WWKЭU>f3irs>|_s%L 7(ցVձ-ˌu7? c"` T^HHG!3FyeL[Bpb<' *<8 D$ܒ#rURoɤPfBz)yGzK u>R! Τfo$BJd%$a % d! R 9Ez#:(D$YR  ܶ\01i!(΂(c>Kې<i@P)PuHm{$|EZn2ksZ*NTJ"orIj@dA8\, -sYINC(ׇAo HGfY"x YJ𻓞80m*^SÓsh yܦĬhdȼ-uҒMBkznl տ)& 9O9Y5 :S H]ҝ /ș-#3e۔R=Sap[/[@9 c][@KAdqfh4m65;DJ-`!ԉ"M4 -Dox"V1Q8R 7Nw160]Z AJ@7 &T.DxP2#h-mm"a 1t6l2r֗R+;+$n/Zb* ,=W{Sz_5foI!iP1 gtj*QD"iTJ͖zCErY#T6r:᭶&H]Ҟ-U1%(T۰7z ^ig9 Tk~mo[lc &Z)#ֳ|&RPet})u(79h֡\7#*Od; fMZb{Θz\x[wj=F4yX _hi&CsD Ds^Mhsk-x!{1uW; O E "$d%V(A:U" c,KHて)Ѫf;cQiTP$#z4eF1.@)Eo3VSx:;Z* ]Eu,n[ڙ/h&=LkT3JR9M@)CZcƠSrFR7CgflJi4vxT1(mr_jY˱:\YM _{hF24OfwFsT^3A%^6`d1t~RbV$; KB*ЅG"A;2%1m#օFG#*k[!VZ{A2f4%D*@s4 $v|PF< k@s_RţW[E3  JZ+i)i{-a BGhZUE~Ji6Ia J'9HX( B Nh &iՙcqs`X<GOO,vz&1=2<13GH% /4*2NƺGFБB HnT64f [})^ #\Z +M`1]8eց@sh|]{; a׌ii|5 {VXo&R`j0J< "etf I2h ʍINܰegaT[&@4sB}H!T{'ЍDWY !F~ 5+HIkHcBlA/gX kBądYLh.0 0Ŏћq9bSE$ūts7ea[_1qkp+U) n/mzqL8R aPNXtNr߱qegm ~"u_eqn+ErE%oCG߼9o!-2( ם;~f [FWɵ)5MkޞÌk}zCuEl2F0=Gou==Җk.."8x AՐu`X=%k,2 bay'z3Bav][WBenEA[\\s82ędG_Q"Gh1ʇ.2~qr~\|n~}Nø6W1%+tVx0i4=bVJ"VKȩLL/{~BG'>?zwo?{tB>9;?{DU[V[=9cKokltW&\%0&Jh!fŅӸT"wH7g-γmlNÔ YUpv83F޽6O^%p gmED<1"2um~dgO?S:g-24!-R*4,Ф,KSQZ#3MXV.FHvICET!o[nd$#>1B5t5]il'Tx3`FdIܯGQ;g~^x蜀l'JSaqݟ]Fe@Yve,AʾeqЮrC/8EB/ j]Qm$h*qڶޏiy'LE@) KҢ'hUY4$,JbAy#kkZ5  4qß3&H|2B:n_:ݺ[3^F=xcD/Y>bş.z.aBn;(&x_1bMͿF"WZv9j[ٻrW<$Mig@$yKj#俧%˔tdmQT._u6r<]i8y;݅g+qן:hz$5m/m0Sc; -ȍ8Z9(Q#bDv78r(΁PĜQȢcj]ot5H:{.:[`o Q[LP !U5FU^JeEo u9n"4Tyy>ݷ*/}iz - OVB\!f4n""@I?m ҆E.B ,5Dm!uE}%h.d@F' .m@zֽ&ӡ]L} 6Mɾܔes%^Zp1RnTn)_ 'Sr4KNz)e3r:V^v~r|7>x qjמm>ख़3 w]"w6y~"n}B~ODbo{a.hh{zZlϤ10v|Bh$[rkiw-|YW#wt21݇܅'8{&N>Y,7G.SXWCޟtuz]ZюcigQ^{z,~Ĭ %r:} eTh, , #w,X!"p.z`UqшT@#6\_W8i=eв$ioDB f>|qq rˆE>xe%ڠ* ٝ'nDEkL+ȭm?ˠ+o]Iˡu>c7wq+4W֩ۑqe~viZvF^վUe rK9ZKZmG\m|?#?m9 E#':lƑ/41 \B.{g$sA0'?xփn3 'A&côZf}DȤ7a?D6-t\KA/-'-Hq"]5J+:;2Mw,̌#4W{n\ObRĞˁG&iY7bU`u.ѡZz2 r2[ɲ&G5#c/杶PiXW"Z}6AW CpfPlq86:8ip6߽T )ŀɀڍ)zR&& o4ΑbP:ouXțSOfqPǡno|b aSuyZe~j!Lȍ'b+_JxaߓJ%#rI4RU%W㥨J}WWDvPWP]'v` bU%/F]UjT{TW&zw3.|?8k'?qP&_-#>Lykt}ٺq-l{VňMoGkKmTr>sW=6ڤ"u_>~D\*!'u~44>U7Ҙ>>ҿ+X,z;3׿.wC`20<1gh-xWd |M~i;+F gR;͜OK_RDEQqW!Vܥ+`<"X+ͥx4ZT*<7[?qί^F9up^u"j0QW/RzK@]A]5`'.G]UruU=Kdup] ک"r*RtՕ`bU% o^F}WWJguƱu^Yo_|;UR + m̧ x,`A%ɣ::iickV!4+(I(mp'4!}*/5K .zO yj7fLQ+1ߏc~?}υ|S~M/I^% UE}T#{msYxfGf1L)ǭA丣Oj۟yk}_:UZS[*M *,>aP<[ QrIҸ\b@(&tT\"dlRnX%]R&u+ 4bn&($}lf&:7={3yPM!ʷԌDF6eU6ew.4mۯʷeW7Q\Oԥ犐@|(ijp` 6B,DnrTp$H<Y9aIb1A#IH򤍀2.!H%ryҀ8}K’6 q\(!Yrdc9kFNG9՟}ۯ`Z*E1 Sr;C5/%ZGٹ*{G9!KEAO1(edQ˔ %hIN*Ax7OZ oku"k%^dQONZH&8:\=4XBQ\Xd^GU c X~t4CC7AΪa~t,5t"Cו$u&ץ[L΁D. HjJ)_XS&ˍ"bMxv5ߐ*e#Z$g[]2ExF$x U|?{(iӫgP^!RcgDXG y\25D+;Jˁ<njb)-GƋ85 rщPv@L E*Ll mRȿQ:/JSs>y\!p=[%mvR3J 'άu TG<^`Ŏ/2'E3T8lI+V^QQ"dܧmh*qhY`92Cp1 (<%gJ @w*^6#[˝~ r|7~;ن"xwEכj׶0{~ t\tלO=YAziO=Ϲؓ}"w\cM>'"^TGgvh4tj> wm7ݵOf~ܢKj捑a:oKqv𐍺螉_&GnӼ;5֬zЮ'|umQ?/?_jozno~&\H%l_NUJ.^LJq}/QtCXf_4?J|ֆ\=[v?12,זa iO٩44>s< İI/Mi(ZP Z95\V:Evc lUѓu88:{AP _T\&g堲eSvxU_#corh_Z>g3WQFGυG8D; D%4BhShjMVajUiww#=3\ o(/*uY Ed^h,s ֣_|h iVlh;M'Q&gD\p8 ŕP[M;C)kt1q!Zp`2Qf$C@'Ḍ{Szt%AWT}s)pO&w1]w?ܼ;AHs2eڲg֮Kwop*0 NnRne4Xq?[#Exo ɢW<\LQ#<\>\V f'%Ϻ:ٱ`uBBqd/6L ,jdSQ2m{YaDdmqYn]W[wZnRtU/__C_OP^⌎LS~{&f\xecq2WܕaO:cQ LҴoĬA]bn%5a=1gk6o`1ţy]`3m}m+ ‚ MA@lPԸ[q؞8i!6r*%Td #,- z!ASHn AH݋V.f }ړ5h۩nOiͮ[M{v֊jS)wSTk!PP|~C(fD 9Y>,q[GT ˹R?|ɿv8w L[P'rpdzNs||%FydnvkZ7hL?9]&Zoքa:}Z4B\lO E "$d%COE G$E,Cx)DeU;QtTP$HBS(t"8-حhYrpCw[LJ ZMSk(#SRɵwLk'"rbhڡhvYhU3JR9{HC┡^{-THƱ`c)U1H*TEqmkyX3ޏRv [MuhYB 13u\1 䚛I~7O~}AiU7|>5OfFsѸT^3A%Ĉ6`d1t>RN|N KB*B@]&ϣBa Q !.$t Us_'Ϗcߔs$rVQ Ɯx+'Jr`sr5NH`ɧTCb(l1WUc..lH3zizJzu298YwG}׺ գ%WG\8+7PP[4/aFX~|[rr6=f ĜOANfnEj.'6Qo`>8x 4$HH Vt4 kFa bap,fK>g=wޛp7iԦg՚NU8<@p/ǨR#'zbM&=?8m9_pAxq=Mb9KWeSx0Js4W>DϊViW Q%r#.PH>|~߼@o>k2>D' 7??x8lhaeO=5U#7q ap3? F8?}I jK,iTqrQQ& G(j%gE<ĜWuW;>Fݹgh q LѲR`K4ouKk6܋F-B]>Br:&|IR1pjBƀm41:`8v)0ǃ@$ &g&eF2)P|W ³*=J-AG^YAQ_Q*ށ.0"8ƀi qv)SnKdw=y $XpG?)K|eR31WMWG5OP~×[;xS^!!u$<-ier 9!o/SL 2}JJ~!|)5>+nL`U]Io#+Ɯfis nӜ3üFA]R~n70})ɖdeˮ,()Ɍ/6ƒ: {R:\V. ǗǕ(u,YQ\p^g^GS븤3JZb LJDXyV{DQ*ݶCw"ZVeP+lA) jl9<*=~aIGu/e y?smV\`/*Ϟ X+ì,H\ЧT͒Ӫ삗;!:fVE?GQٸ_AQ1dTtvBh dgQҪoD3ؕuBPAyymSRxx"G"g͖'Ύf~[/K+Z E("" VRDA!$6 cv,y,b{w6tɉbV:YfUPmPZGtdKV3U+-&4I:k?XߤDY&ϟs1#xijKH>(DJadMQnF #EVd7`\`,H19!Z*%b :diG_fCnjuz~,kF"i >8XRa%ѪjŤH')5 Rш&SpG;x)c]=t4ע6F?U89ËҭwL8ֲRwǥezE֐Z$r/,`&y:?HEل7Kp $PH$전ԡ RԬl쒕PnXߩS{}XC*&(VפRQ8YKARU8KE\R4iH{?I8\~ηWz~0+}zL_OQ|})qhdMQ^x2e1$IUD&ct m Li^6[Ξ,3(o;^T\77U=?{Lx4p;_>sc*+1c˫E?ۓ/.r*VEhuo~~To!xtut}z|Ǥwcval~BnEOm{|?x=fjCW;y3o1}v;|t[5Hs}~"KзٰFlsPֈ5~PΞ}/l`S1IbPY)(ky{}|2濧ߦ7Ӂ4{Ƽg t'?&26MޡO$pǟS?OZ0z{pqIo-IeAڶMSt ?ZdiLī9F0o 4]xB#R֊|1REv~Y8kNur2aPiM.+C5Ø:aUI>WD0ɼ~[CWAnٸ識+ugځ)@fA"U* ^Y$|E+CTWNuew]^gr?ivqXۮn^}8clEݰf\I=Bt.G0l͈^*G<l8ƺ>9a!uOH؝0Giq s#zDX#(Oh` }W SEEHV/p׉/m*W/̶1qG.ӽ.#E$,5m5rwt~5߲1s¯^J|lqHS8qUf+=/3WU݈u ګJ&1y?fQ)%jrT8{VXObi*m뺈ɨ}VLZmTg*E<_]ҷn`U[̮&E~vF&Į(Rh t{ ь!#rIK4}Mlp#!$ƨ2 $dE#9P[> os0eR= ?@>AdB9frtbA5tTIR?Jůrvvq6fLk)Ny;;:Q;$¨@}jz`\-TcqlddJled5!a,Jx&XR[/MH@V u{=B-mձч יf鯂oW\p-{My!8>ѾdXk@^_5q@ 謥<;PT;<D3;۝ΎY x@ S@J!-#+R šʜ=j!,OKYo=\=C;D}bګzb &2; cIh%!Ig}?Cҙ`MJ:g6,LAc5g?Ά3 _NŊҩc4"hp]NӴwBlJ5`u\m ãS?%9~BGޓ sDȚe RcN1H+R,ct Cegܺ:'_$S) Ʌ U2:$OJDU@ !,]Sp2(JHdLQ%q6{MFOw͖t>OȝUg}|u{4YKg-M>.yVR6HEotpʝc^(!,Ms E2 pdI9k>qI mGڙbѮH妜zWri>-J>yr rKjQJZsfk) Qkyi a/4S`{?d .c*wǬ[ᭃ_BvpQRB*w9z rGKR[J.G02WQ c8;l D*A 1` (Pd J2| KIBDK!,,&m2H$'oh!F=jmfR&8]%f3Ώ&'iԼ>p#b"u,Zy病VwQ3xx,f?XM! Fm BVy䓆LPJ)Ugӌ1"aСPTԻJJAB&?5nC] YEۜyj-^%D(>I;)#"6lO.s{){b=)-C7u璕s @$% "QHxeSYa$w!4Vi !`fʨe-\E61F&6#TYo$y;-^s>8rW}iA?&trw?q*aOI^ZS5W|#rN˛> ͈9ɿ cVU0̯bR}0.SJ{*/Oi<2&k^Tz6sJ()J;4'34 "0eP$a&m\rQwͷ!A;=RMuRv1S>'S=ug*dK[D_-_mi 9M{\ܭgvw6`Y񵞔~WkmfEȗY-@>Nh- ^m5Jr3\dJ${vW;pf8)9Ν1UOw&g7#~,uN|*o_TN敳KOdn9GZqyɅ& .a(Y< 3g1~f:ѣ>M2^nZsqI2;>? >7}=BJ?G3lC,ltnK5gS6/_Ç0iLRϮ[,A[ ӯtpIzowԙ.BW [2Tdvw.PHӳ'>}?Oߟo<=L|rv=b7^H1 nD x'aߴkae&|~49~ׯ! n_B4OSnc"ɝ6ҮNeuql4tq[>rᄟm}1ԟvHĈ@2`{ew]<\}'m(M=sh / FPnjR9a-g&e)\"u"aQ! [cCB& (Q:mSc+4OD "HF|@e݄jhjj6OEXy.I<.u[Q{~l6ޠsM|b_(vn|!b{\ c.(\ԶphNCV Z1j z'P7BZT M1KhXj$ĂF(9[crVI*x֗vEϜ1Psx~- \Y8n#IⰃq\| 5QR&s FRK<Pf2/dٷ*qjEW)^vm{jXŶմvi PM$X0z052@BV}mMXS!{9 B)W\*2`}L%W-\BT*+noA_|yUDMT̒*{ЏOjUAZBHzFLw:tLRutHH)y \0 yQG>*:;9?-c#>?ї~SJ_7'\ JP\Xdp FAfVD^2(޾&^fE+/"T'p|"Ŝz,nvEͮ#9gkv0 `0ZݢTByR= B>Rs(pWzW4Ezp8rRzh{qTr_pWz8 w u(p2N2p )a> b^U& \ejwT++Υ|j&?|J߽.dw~,r@Tї)o x›@B'! NQDmSQyx܎OSgKA2/E,Si˕CVriNn2@uT}[< iz9y[E1K1`t Z$ѽet{c(hǩj%ŧO{MQQJ05Kqf zFӿ#L\4)rFm{>d;nKR)u'xb#u&,A Ǚ-$%L_шyB<7Ϝ?3E5"'@Y8$ Pc $ƉLL7O$.}>7&h頙&lR*DS&e6G~,!ֳ3#guZ5d>28w 5Z) o#[%꧟ZQ<x1? ;ZmZ,R?2 *O ; p0L:h}"W5ٚ;)e o2/uG43C8uF J=W%&2` ܙ!L&\A!T nV<ԝfNyq/פ;iyc__'?ZPCf_ܹ/T6o·n,S`D4ꭳFz-6.O+ݮ[oeq-yR7=T3_&򏩀v]rNQ淸PT-' [m}f{#2հw%b-]?^3AS{U1|&'U9*e4a=*c,hWMC03'0GH% /4*2N:%@:RIqT6"y.4JQ#et8Nj"5^  awUc-6I@OO;`,S.8ǺjQ(5NY븈H.Iƕ&Ms0 .le2@940 C;uP(G9ir2L\TbW'J3P oTDD%YО\}N^Xy, <3 _!&( "RFga +\p@+$:)kZ nZ$. gSL^ιr!gk6zhF,p"F~ 5+HOldB!8oA߷z3p3?<.Iq>;㮻0q0>5|#rǝ?08$oSsea]Y{}0'wz3y<curSyp2- g\گ~u/&s9\'?^Bz'{wdm7rbPb`żDn^vym,u:Ⱥ^ƪ54ʻ q*#c`s#įczD;6F&T~8er}?||,K>슿y)L0Jh0,m{.BW [2Tdvw.PHӳ'>}?Oߟo<=L|rvٻF#r Ypg H/>F?%eBr%W=|8%6ERj3=35=]Uz2ջ.8L_w"٢ćcSEj~JuՉR#wȪ|vZj20FÒ%‚I*6"PeshR,I+Ybr+jErcWtT.gIh :򿋜X Y1rKֻknXxX HI U 8MQ:kwY$Dc'h ٠P%Y浬BNxa'I 44 9gV>MT**_j'.8֊%6q.GxHz,e@@{ƃc6` TQDpUŒbkT,x5M)tR撳@2IdmrMpI0) ˅{ʺYpGi˵=>+D|0-JwmQP?Z*)J{K>!xURpXg{JTi|YEr=ʱ5}CVsC7A?bc4fjTkD=$`hnPů/TNKNBůL^+Sid+~|Z@[AA-By36|LlhLd|H[֌BNWM(B\/Q}4qK*:R8'aES1K hTIC 1 eE2$PB7$+)yr_:VMl>>Gk/D|z|T@1R>>ȁO KO?q7hjr@ GXj%TxyLCR^!5$8 DAKa%Ss`頂&Ć g^%Rp=fi&ruPHY"!r,YíN89[,ї6Iobp veڜmBx-n52OM&Q?J{ gj[ϴ^O{vzi=-I:k[~wH1^w4nptyt38#7]yO/`ϤC6ї-j6ZQ'FᄼHx[JM[#մ7Vh8vzHdU&WSWZ.Vqŕ69?qTUVcW%ZWhOI\c#'#2KU\\ejT\!*^2[~zaQ)?14rӨUWOҐW TqOψ U&ؐWH.TUcWJz#V28q}*S)EWoQ\p/"͕ϟ(/h ͇3,Em@GYAYo)ͽ&/sZ 5DB$&xA%MA&r ӹhȇy e0,l2Q~ӵ1n(1gy~u㔿!\G;xx)a-ѥ/ Bj/ upkʹ_H/ Bj/ Bjj'~!_HR~!j/GR~!_W`1O} P<8K$X88L4g5GgTjjƕ_H-\~!_HR~!_HY d_H_HR~!_HR~!_HR~!_lsIRERڏrhs3_k4hy8)+rcLmM!9ܳytj(C̉P+&IPQA*#Iv[Q*4TD0hf Jq&hf6Gb%ޏbHbH"D]GMVV-¼3vLﰜMkHu}-yԪ iESc7́ ~YAF <\򢀡#R-,m"Ԁ'q™&1TQG|+h)z,zWAf}:?sE淜ǧ'谯y)WfGs*)5khy:%D P)xۿ'N%&Q֥&-\˸Tdr?R^Nn/7qYeޢNZzݿ'}xeק~>4Lokb[ @H}٭E.ZƐ]5|V>᭛{ϟ0g2.C2"w%=]* ٴ-c-4|zzP %(jcDg8RP($D)#ȊhsNÓ_f܁P5OZң8X8н [Յ=PxֆO38>k6%0P(tN) m$M#[-7uaRuwgED=hD|lĽ)p8@(,qTxѐPqN 2@t4EMi"B31RH. iYm\D"Fn2h|¦t1r6[ޤ -"̔}[+|Sk-gfvy=m"㕶Ԙ&9eoy-WF.}%FriM&04AnҞd(wq1B,_Wu! NAvMSVa2xr@i4}%[Yf(!G̫wpp1_ti:+#g?d۬mϪ54ɧ q.#yb'R/{ď!퍚c[]ZBrۋax՟t1?ŭ}KV(30+=\)ߟIiW [9#}B!~p7^x2}//?yO^m$pؚ߶"ܘ?^|c˩q\SˢO=[ UP[y ap4*D#h ΧqsA䗭(u̖T5J$zЬ 's0|`7`!~VD4y # +~nu֏/_J6xp9KKC[ 5R*'#,Ф,Kĩ(rQwޑ&jQ}y1!h!pB޶Xϩ'4OD "HF|@0B:ګxNћW אŸQԞX[^]{u $J}TBz_7dZW[ReVNӮNhW] v%Ȥ^Dh FHhHVIB,(obeK 'h"4Q/n.t9c졥u~Xzz;t@q4IV߂a%B9B e8k$b*F `3! VFZ|6-\]ܮO>ٮk{ F?#kvH끐CX-V q^FOg -87dbgڦMZ_{䍝|8O}]</#9f)G(iq;r@{Z_߾ڎZcvKlE%:S2I8Rgmʣ#|u8jX}Mĭá8 &[b{}gȋe|k/niz{|TyͮZ'X'^ߙF> z/߿5XΛč^B{/{ȍxAxgߝ_b0xD*,ɣQW\EZn2wGTW9!NM6?\ }xS=?@\_&{ Tזdǎ/jmv64*zɞY7F0a6pC| jr>ߝtA ,Z1itd)HҢ߾×5g4FwbJҥymg\ip9MKP TKdK27~a?DZO'(ζ\KΣY aWc9@Q۩/bGm >qx.}֙~:]i D` Uq]2!9 z G/5qg uvl_F޵0FhT:l!PFT ܘHm (6JUBhkhu+}rjD1V(Hh!j#$ u. K@DI#3B.'ih~$Α~ KWNyȒ. :rB cE0*'钊S_z1r6h6a*T,E[$&=ZK}Wo-t]ITɜ 00b r?O^PF*% hU gAv[LrY`ӓNΎp',wȵѰc Gy<(­-䒢rfI\1lI#Ր{*;YRp=KBIǨP]^XΊ_.i}וJQn%41F!%ym$$=\Dk@p;D|y*?S4"$"I$UXrXa 0 Dm5DmCAbPOI5&%oI

켎=i×ӯ>il/[nm@9+77JfWwOCYfZ(\{W $LZ{-v4RT\Ou42QǸָDK VH ﮆ|[Af/ٷʿ^u-˫jSZAm.sex mIB_62R?_CKrAOD*$e[9"% IM0$qf8]]]ﮮzR#Ԭ6B͓05$(3DgT~RϥJfpx.H3}) R])o).=_[Β9jiןG&hD+m%)a FΣ~y~xo7 ,Pʁ͝I$+QA*#I.K$.}b"91:0A)΄DU s~Kgۈ2^e rЍO-OUEOi ;7{>lr>%Jy돴9Ϝҙ<3 6:#sQaީO;M R`5N>Coql`bӽtCܲ794#P%Z3+a`6 lSb&ù~<ۆaascŝ;M"Uh9ؐf)xK*PʖXهEfIlc$([|T Oqn^8pݹ&2yi Ik>(B[U $ hD -D+S<54X82'cPOBA|<*8e5G]=\aN~N?ZV ONhZ:+_j6u1i%hrz\ozP!y/ )R>"ALw\6 &eF2hDz%S"yzL~}sЈ܌|n6؏shGєh_&Zn"}q8)2̂,㛽䛝)An2~稽A^Oc0'u-[VP+j8y N F"^ BET{F~x~>l{F^^4ś/HrvF&O{Ӓ~ S5pAy,wIyS"QGz:Qy|joe|]&y ^l>Ct/T,K?D88FBҠb@6&(浠MXDJ)fLΣ*Ua:IUDs 2o&fU/ILFd= y8a:FSBNM:7qy aJxheUȊ]aX ~af}uESX.N.vrqEÅdz޺SҳX;? m6e{}W [ )/g%ыGKWAs}CS8wM(AOmD]q32FצTM9>4}<Ìy[rr:j8#snrk.?_fA>8yujsK m$AZGqHmðalfX##82 15ppv5CzyjrQglmZsqMIg<H>|~W=a qKX}?8k9ˏ?9j[s(Z4EbAWz!r|v.Pc+*Ro4{qLͫo<׻=c񫿿:~Gu%@5m-5yǖCx Eg=E*(9qqc\C\ڟZh!fŅݸ\e+}ek~|6[6aJ= Y/p83F_@ vmc%p gmED<1",2`ȎO/wJ߶Sǃ4Y%X&4AR]JH18 4)KyTdΩ{LV5=?n xZȤ!"*CHۖ956YBA: d4MBP{ѭM<({s9`M}Ò_YnEyxkՍI;l.i8BW+32tl+zE?Svj֋pP|yФSG^W˄㥟ptRu\ [ARwΒy_,j<2I8BGQ|V+%pv/N9x$N0 uR^*mD)袭%!޻C. ?ۨ#_ ha ⽹nXˆ׋/ͥ/kk^bz:u6lfw{M+~gr9]|yl>{Ӎ':{WA ?sw`XzݼP j7JI*/z2V Em+'AFSӮ!bH^D @!-z2&P%@CR$!7B#ur)\ІH;,1AxG:UiW(Y)pֻu:;}:+|rlq~U3DΜBYGoOb$l88K0I&2ٻ޺n$W}i`ť I2-DKNiS=-u^ZbXuC('ʨhAFAF-6οqb(벗@/] cmuSz~Ċ}8lS/}k@M?9b4bgoSwuo~:i7oA{mӧstAץu݂.OVWBZޅspΓ'JJ-8dYV)8Zd(;]k @Y_]K,Y1pKpԉin)TN R94zމpO;i.zUv"^'DXVv"<==kQ;'#*ٓQNpm!co*,5~M_G|#E&Zw.X+Hzt;#bu_7"{?k<^8*\g| YWyj^bM,MhU9DPMBT3dZ-<1AJ^߹5?L_sg r}rMPíkHxܽxb?8MaMJ5u͵q[0[uV3%*]9it{zkҕKF'`qCJGG1.N+G&rr:%œIgvoo v62\Rp!`Qzwoиpurz$t6yaҲ7b`wbw1ts'Gk1`;{sN  O&,s{W'7Q;M&'ukvoųnF~0X '#qDra͎z?-3iâo륬.^k5+-A?_l^ΑIdZRφ5[k\]h[v}I*]un:[[1NT7Bڃf{dn}W={7{}p3s){3J%=|K(vSP%egM 鬜%p)K'R;:ƈa4BfBQG"Gp҈K*~0iU 1Wmϛ: Y(/Ȓdc)?TP)}{ޜ5A:^yr}4(%5VæHr#钍5n:ƾ3:%aM x7'/{scO1QQ ~ȷ2s&1$KsOkh]VI!5*dlzaCKU+ܜ-2f %xD]E=%Ġ:*D{|:C|줍] P36:)4BAmQxith(*u :݁-B X 4sX?# %eâaDy=7XuƮ" bbu2Bck(|-sb Nul]z\@QZ6f}G)*:P|;zх6P[ r FP:g,.baIc[5lrwUXF&Ed6 f#>I:-il5VԀucxP +Z#$d Bf뭋H zIc- "( AU.1a xk)& 1<`pƎ\"1aLNL t(R!% yc\ T&/3Y {S4,N) J8 seԸYcj ,(l NZGBR! ƺ]R{fSQv/J*"z_|£YcIL'@H eFiu":(5K PeN-2l=x_ZU/<d3.g~º^,KTA 1cHQR2pD(=W.#d3s7þiךl9#JhHWo#fVX94PQmA{[:Tm O>y5 >«ځi#*`)o"02jP/FCׅ̆Դ~Z)Q 2`^ F(0qKކRZz ҭ_;"@$\VycEc6\E*ҘD 02P HhҙR+>FRClO]Xv`Y1Z`4rM3]Ob!wԢ%蕲7 0@56[x#Gb/n9>>t*WhE/]SGUڋ-gK߷uP8kObгR1lK;Y6'L{| 8ZUJtvv(`w[>y%LF@P dxňH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D l@Fp-Ri{dN =y%[(YQH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(0P6)ȺG D{k@֩'^9*(w%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J Q*B&%W&QF {J XiI@Q ur'(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J Q)>ZO)ݺ>wv:⭦i{}~7{;YߩN!.oujkK6'/\)p9tu@ ʺuJuptp]~a7'wYo^{wN.O_^]YeXMRk*F*pU?jNO E2߽׃蛳xax|ɧlU:kSv m5i`u;33`^>G!TK%Q$8 M}g>/_w8r^c3&kn0Y W]*좡A݋?0}+^?n,OE`;wB!mMێ ֬7 u . xp ']D э{p 2jOV>we Ɂo.G1gŗs$tYmN~U+cjC+Zkfųڃ%Myo14cʜoQIRRT4 "j"AlyUJ%?GgY68bfa30fړþ}Ҧ+DKN ޵#E/{\|_C6;{3n0|ښȒG-_d%٦, nYdG&vAm0"H*RHZ9LX^oTԷ diNeX]۲0Av'p󶇭͐0t0 f_9ż,2{7eo:ldtoznem'yoww xwHM<䗥[;8wg?6+Zˁ?9Vx› @J'!*NG&hϓBxHƀf Ǭ9Yﴈ!D-P냻' L5j$aŁ>BTvi~)s,C@+ 1JQğ8"lцX/_1w1 j֏'zuN6n_ T. -nz_PDշ. fhPqzh]#83mՎC3k7{cόy8Nꕥ̓yG~Z^Q=*JN~R_Nn:9I),r |' wsix7M~99GZhaNޭlsxg!֝\\6w'$N0'wu'-ru~d♅va{nZXz_UdSns&m ez#FU8)i;K$ ճS:-3;74oSi215rI--JOfz3n ?El͜7<0I`oy콌a0ls X=c]7q`nha6x0uVu&wy_3qu (ʋ ok߀\Qy7U? ;RQ&(]v5kr;uݖpⷡBcosگ:)"O\qxr\23BV Ϥ wvOU|B4' zo6yX;C6v/x9rf M'nϣlQU+'TTK+`Kyg<(Ý^cW?~w$GꪜRhy8)_4"trZi'툛G?jb9ACe҄$T$ GMwA6:D&;糈:qe;nڜ ē ӲTη Mhs{5(Mxҙ<3 +iD>28w 5Z) o$0k>&o6yťoZ|ؙ`͋]BɿّЙ/I^٢ Ю)g|%$g\*c53 rQʡ ZD)KgkV;<0(wh ۽e(`0>D+vol;OXT`%_LYLZV5TzYL}e%#Gc% zPb%ЙtUbUIPy0È#ީrAat(C;Q>`߽p3xaŔv8PPI9|$Bz%+ e ' Ȋi+U:V?ā~~~]RJxD!lJgD#F󢌰Fg:iI,΅sa=76py$2AB,<#:j>1cJ*Sb?^AA됶|~R3 ܑ"69 aW_Uz@wF7dǺ!u4;D*킭`!ԉ. [NN)WXIk@/ǘcroT)Pxc ]&FKR,!őypUWK_lM7+!3tiCJ}tr9MOZާ|c^0)$ *cU04vФqԕ4C% J,eĤ%nUF=FV)kf@DP rg%IM.z-);E?ʼn4PQF kN E+h %Cj34,PT3JR9S'zP!ǂAVX-,/[3nR qcu( Ӆ/,v3uʖrߢNc믵olYArMCt(~*@q/\@]x 2KeNd,PAAeqѕ <NC60ceɥ *!#$&DҥJ[c0SŸZ[ڲڝ 8P͸ PS^S-50Rh z>LB*!gHQQ L{#(`]H)=l bև _t͊T4b1xFT54bq<ͽ&/˼OFM$1דE?HDtA&8m8@^" ;&gJrN?uY=A%xu[֒D YG'$:}~FzyEh +ظ}8c9q0_u۬[]XMQfaWz8G,q-nҮH0e9!=߿D!/wۏ?GwyoQw $TO&' +]c=EG==UPr-N0jчx#ж.]ލ{HY;_#=%):4 8PQ z,,C7ZBi\:X.^PqL/8+ɋ).eԋg*`qqI'I &0т焒qW@G#?{XJPs:Eh he^(I%ն֋K'_ȱPI9Ȳ,%4^srOZ)¢r [_"U~@2VYrB"$[t)ԶXH<'Q]?9i^7|gfvKs+{rqht@P]^'R][_[֛otfKe<-L/wiz=ܼjŦ-ng{]NsJ8SoY./"xf5/o'cl% ۗjaB=6;NaxIG r0Ā`0ĨE& Ntg01K/Ϙi:?\=*_xi.,Yqtٛ"#(MG .&L`ɨ҂m>cn[_m̶WA @8!u׆_n>?!zP7x{Q:=U5=9e*Ԉ뭐>08 v@)༵^8gX'\o%VH1e]Z緲z=}ώQFULH%I%#02Ѓ R9Pt jurXFBQNDQ$2tVL0p<)$&r DJ+&㨈F7B4"]̴ )Iڠ Ͳ)6VX4,aeN24}/RT5t|,l/1ǏEh9#y߰8W?56^nj&hA\AxQ[)WIK@ Ӭ/,X2*RZ>׮~$ Z1h @Òk' )r9gH'*LD]<=WyH>!#vfn*G^d d7Ū,Q,[):aL bc6]Q<{gB2K XdRVeK,IN$`/b_@PB1m4O!]u*D(k&D@/L>`h]t TždFQ.ͤnc a-p`h۠Z{vquvd#BW7z<VZKxuJRB&c\RLH:B(Rt:T¾_L?8#}i=۩r2uUL̘=an WAߺq=?b|]~`pC&]d㯙ʗPw݄wcϖɼkێ%R]_AXibR$x)tF, B10DNlʼ 79 )EA*MVFND@IQ&آ{mm,c9)H -J{2‘8C^3 !5s1K jWBڽ6F1ۈ|h5}V)ǎR e)%HlJ"OEa"T0xPP(M_Hwuvi_l;MwfLoɧx|8?E|b謰;,:xh9Ezyt67_@gy9]1!ba2xbܓ &P ÀJ!9Agt߆8[K>^2H<]t/{[묻}U/'1<@G"\~65K x}7P&opN Db0¿I"$T,OQeR R T͐qRr( "PZKPJ|$ֱ{,ud9 hB͆=zkic>pq@Gy|=[yzpBp{3qv:+;9"ACQ(ff;X#Uqx].$J>[ { !3}αr*-7G'L`ƨ5tq )M4Os]2CzY9h#fb11 ]:-YZ.b(bϴ 1O^] d/a!CV\:|Y@' UI"4f.7 ǣIY_~wdkǜSr_=yFYվ2O||hЄ)C,c<cuB |^eVvuݡ1Jcs( SQN)>l΁.NMB^"f Ey[d Zfg=yr:/t쿔1ٱf|<Axj$i*`s2 YWd"I1@A>,29ض(g`4,ŧ̞ZW$2F J+P)Z+\jj8E4_Ib+Q-v5ɩsLFߦd99׷Hd/ovnsӔOgXkѾ*gM 8DRl2Xc("8JKF[["[mb l!5ՌBTIl%UON.9m쬓+ F)5Z36v͸D6[Յ.Į ΫdD|;kHf7o ݎɟ5klI9IG2c`@4d.бRV>6,Ω< B T& Yd {3zSR."Y:,pkx{C+hfZ4ڦk`7oKtQɠtdQ.%%'wum%HƇ׻ɦjV<ܸ\@K"J?CRidt4QA S |.~񰠱m  !33^4lkRAf0qd_ NuWȽ[~s boTD=#q@Ui")Xu_\z#r_ 'أK颅rWDN a,oŰ,8S$aiuvLU (:X6E zFYǙud\.^goZT\=pqŕ;ZYhI&JݨEYZ xQEb-- 8ŝK;vXp Dڬ[F ܋نH G ᐯW;X(5|qxc3K<(HJ:BAj"2#މ||[d]/i+{<݁[VG¼O| f "J$t;(6+-ٍ!Z4\QL IS$1,O7>m&5|Ţ.)ۖpfXHY)-[Dlh-;!cU(Vܣ<ɳ$mX&JR(xkAf$ <`͞,/85k1h UuBҴ!`|Ђ$餕.)ZB,: 65)pF%بrTe0~ )[bbj[a 1rȈX,EDt#3`cȎE1o*č7Q=I?v\>0J|f(^OlEYWwDM1KȌ$?X 60=LG[.+SaKxxO 8Li?ұ3P7Bj5ZKMa:ͧA f~m3"`* $D!#hIX;uDQ+OȘ> }Z'U|)Ǜ,Fa4NutZ  {wxyz~v[gk|5mԊ\ЭyBۛ[\|R8sO֣YϮ^xa<*w#im<)'j;KEj_.ۼfrO DH(:G|xӮaT0t,S3Xggގ9>_9;GUb^xI!/l>9Ք;~~{ HMq@j̖u;hHK_#uIFJH` /vzz3͗W;ߗN<)^k{<Ύ'yytkPjaz&_Q*{S)ŵK*-|*UC7X){o#b-U•|^'\}3p飗Rhde\=J`qZInU+[p%Wr>z%)GpUboR \Ui-:\U) •NHGpU7pUE/pU5jJiWo_n]W,aBBk;6wjqDTRPMZ:S72KIAHQ|H  eC@ 5BR"g=\nR@fTh5-L1&GB ,*dW]I hk$doDxsR q*{pVqڗJ/,N`[vzʳ&٢YO8/SCv mύ%6P<5QdYP"Yc[1/0c۪whu̴```Q(eb2&`pQDF-Qփe9m@1HJ;稨-(kE/q22B7w)[]s;BKM]r_<ڵ dwcn  C0H0GPLma˕m$\@/C]k5>j@fNqk^J|{%lw^K zs [_b%dP'g0e4(\ R2ق$+Z]WQhb{g6%+-6w+kߨM7=pܦ%srzmUo7_>\'}IߎʸN5'I3%'I?7֩JD%(5̓2K$OeoJc >& 53DNF+%z'};?5MJCKs/yd5U_bƂHzen/MS'4WԻ҂ |wF姞J24m@ jA9àNi(|fJ)+퉲7^B "jV51’Ir5_s6hHo4-g-@tЀ)Ѹu1JYKi#e`k ƾ3%zFY1t^e2.3h)3rdM)qlÇ;ySIYҧ/:*1 мrwR15 ۛ6j.4,xC^c+bA8*VGUJdT t$ rY%hjƜa`*rZ/!RW\Ζ?f-=Yolhg=/ E˻:X+}Q@t{C ^PK! '1!Ő,M˷s_Y1ʠ!+:iL2%3dDZ0 2K T7E'4I2QjP6(:I)gmc-L.WMr! p;7FwSe2-)$m@RBY˲Ȋ)ml)D HP $fqGeKK E"O5>n lPr8iB9 &٠VEUO.ʚS1CX` 9 ћ<:La&մ.N5pmrsqHM^'}, #$?xgHkFB9JQ)X28 J$|1(f *bo6U3MoVz|y<\ލS>vN.ɭ6oy-mx[lĎv hat^f}٦Kan/$%uֽC' ;Ĭכ;{7*gWՓU*ݟtU\7dzV P-fmE`z@_ٗ!C؞@jp=8^;p(qR: +[[.68;9R_ oAg]B*!GDm!0; $ScL|RXE#TƒnRN#P@.4Ĕ;kPH 5,]I|ϯzgÄb1P o_Ei2^,d;LY/V;LJn?`6vֿƄA\A#Z)kB)8KNyH/l2dZT DkX`xK X-l5ZNJIrΒ}BTs -m(%-7+xPHRBN8J@U@M fַ0Wޑ>.tXwmT4~텩Jj2nYEV⎉2oYH=.benjy-4/8\1AڠMg&mfkDt&IZt][s[7+nh4nʃlR5yTm˅̘Z8EJTH$M4p_7}kpQ5|X|XVg/oG 5XA m3:#ܕQ7dzKT:u ]wZo2AA0*1MF{p}6u8Z@Z :Y1`7\jlԒ2$PY- ƀ:@]s]Jxy1${(Dj^XuK.~ڰ6Hu(Hg'ĜQd3x2EZ9Htc3XaaYOGcBG`Uƈw:Oh[w[Yke\0EE5Kj!M2\:Ƚr,խRrB Ke)}9ʘ F,pϭ1&d^f".Ѭ&3(?8_%>);8f7(RCay}seQ6~'uVAh +4b: m{٪0W18:ZPPGl ʱ6}1Xt5'uVpG2E%/1\`#YJɎ.'T_V:MS^Ay0 (;z༷1q,F3/.J =o.Rn]vt%O֙Ai>[N e̤)(N$QyUF@BEy {犌"ې(姘z B c+/ךP&s-q7t6鏓ŦI&%qԔQ{rg/}ݮdm?%: ʀ;^:~ӥjoharZ+ߟӂ#CTTTx%5\%M[je<ڑvmĪQc,Mpt r=)`z`=R \#s =c5qQA|a5x/}!6_sޠd 74Agߖ,3ve,w/.>3<&OF2G G5O *yDC!Y1rZY|ݲ8#22E`{PܦP:( upmBrHmj~/q-<]M;Uenv+GwqBq@2C0H%p -#sCd`33Y*s\r!GH|ML"&(c2PP}\Ma/ےx,~<#Q7<6GfDEYVXV2"ȂU=51b$QpQ#ōg| Ek AtZ8wX#V?$zԞtɖyLjZXh+Eb[3Z (.6䒨x# PjH-+"xC֚Fʧ@/.;kiDZGC={'q ~u}ļ VyrHK/&/oͻd< GFIm4`4ɯ[ FȎFF$zr=ܚvaNOx=6`eŶ6 ? lDg=dy.4ASkJh8@UUϔB]J2_(؀g2@Z’Gm)&YW灕+=?kD]"0t[mcx~\yW6~ rL6W>$V r6'AhCS0AER(`6A6F:30Nץ BFWRAxˡ.b,T K"tOWa8:Ƣ h5YDrdI Y,"{pjB➥qwꛄO.Ff7!s Zc^ǒ=燯S.wS&x~%= n%5-27 θ`Ⓗ8J. 7s?b:#'x"lxoXrHƒw\:0mq -0_/ZƳ">juvb7O}epzGX%Zٟ NhK֠iޥNjOѐkjVK&jjJ559->=>];o}Xz!`Vœq >V]Km_p'dlHm #o6a֕-bLG U,;xe1ٶqT`G]NrӨjך4- ii#e`I׋I(_}5%%#1ĝ_=Vw߮>ǟP,ϘxI_e+*J<+Z+p:˻Te4)ax/o^˛W?~y_^ =Y 3`,ߟ$'{" /! 'q5ӧB^Q(Ajuv*}4F >KPȑ*,ѰX0u+wF?y`qvDzy(lث[Q_6F}٨/[;F}٨/elԗQ_6F}ټvlԗQ_6ب/elԗ/6ب/elԗ{<2Fy ՏJ[m1f#lĘc6bF1f߈11fSFو11f#lĘc6bFو11f#lĘc6bFو11f#lĘ0[f,N_nC'٬#N&YgLdqW(ѐ? B'oB'7B'R D  "󒀿4 k!)0piŘCP{Έ*,8+:¿Dw=_UYo覝faߟ{Ny>ӽcCGAGIkhIY9ZQ-iY':#!~"RןWcoL2oWF,~ F?Fɉo ii{͋#&_^&x!]Ni`}:J  w6! N\uLQQT)b ޱJ\8'5!$g.C9ی"; EnU{2:#:r5>0:e%ȵQɸ_~Րsv*"|,B)MԷaɵd٧FU~xz:ixPx< l |WS/q$eQգ GkIcL8ϫ(qɯ/h>cqGb6kܧcn`1A\yt=ܩs]}ab n"9qm%ѹI:@|mAj*Z+y\RH1\Jlen.EQ5#jnCk\A!.Pk{2ΛTxr^Ot;r`[Ui-dfpdUϵȺȩ&A @9-dYCIAblsyMm~I'vev<Zíry;u["/u>)m5ljcMU7=XPEr!AF>U1dj=4'n+a;{WmcMb C;wЊlM$Z]BHyd|x6/X]:L8QiД =0 J/0Jnʖ5+&*TMpAi]RWBbBlhw+Ect }(YUcbkD%C%gT!F'YBgj͜VzzZ;8rm5\W(|nY?k3A1߈b2~ ˡv&XݔS3zF쁩S_}YϹS^;<ȽZ-4"6^%fȟTYFO2zi3z8/W&s8c\O(#'9mW{sɷj|w?YW;Պ=0qFo:q_{y{wynF| Z0r]u:ՙGٓߝn!r/'V'*&8xG,:8<͜Z 0VY@eU,Ij<GJB"*+q[T; mjE>[bg# A3vQ*9!T?޷9 &^_KcvjZHFuu{F |6k|mД l/1Xd2YUUVeRc +΅]w YR;(9j 0x-G,v7oxdKzqsL+ %(Gb%IBlL.}\ħkNT13yAaC9r+Fܽ'='\wuaf{}z\Ī!;bTXB !1ؠ[@3RN]N40>'c{3B׼!SVШ497q5fB rʎҜKOx}uqqj}W(3٤JbMplePW\ S3VA%Alr"pd**[C%H<ت[%cќMʩ|o͜#scHlXoJsJ6/Ĺnd%ݸ/gmN㯷'&ycoOwk Ksi<&R|^l)R-@ B[s<# O]h]C|z-#)a .)1ǁO#$ޮɧG`\}j8o!#W춨 l$,xaR%ciEUvZ6b"FRY `-l&\$n_[]2=!"ݼpZcG!EzGvwى/AOt|x0^ճ1)Eoj %[)jC2!Đ""!n0rKO~N8]+NsN_߬NlS{{J &hxt7'~w ՞fہ'6i۝箥̎tT#zBO&cqλeĭ6d>AwS<{%H3a&_l$u׎Oբb h$ 9P89 Z !X֕Pփ'vቫkwpyK:2%WJ lt!!f$@:8{Accg}@T?Mv6^8%L3#/=V͹w"KR=Oz.KW-|Y޸yjzY6ϳ̼O0O_kv~Ip%;f..g WJ\}@ (W\Kf-UҨWÝjp%rfU֙UpEl4.`WfW %+1)X \5s-fYp+KpIGb0]5sa1`vpլtf#•#O:Z`3W/樽YlV=m.`ˁfY jz5w+Yv!2{=Cp{n!=ۯU'+cvfn@gn!=yP Cv(ۡ`;lV Cv(` Cv P Cv(ہ mP^]NCEp_`d't40*>ḫ‹:5`{+[W<%Dp|wk}CR/%ṯQp9E@[M1 ;zB|Vsן~_bd$:ōUL AShv[Seb-BWN^9o xPoRk,:VNV&Y T5كs؃Yn-ֿN.58Ǒ;bu@'E*ZR3sZ`[g/٬$?zu ylfqu.&/p hPNP1L7\*<2jn_FE?P@W%ŘYge*bR,%5YBg. }i])Ay8Ezs 1zMT4TwU"JhhAnygJoE}-Ÿ )I{-/\eudAOUT'5Q )e߷Q[S1Q6}KS(N 2QsQ˧FMEX9INqys.(?ȚYr33SIVwK{Gb/QWT>|P%Y+6[9ZӘ YEg٢Aq("eWsnQe5ĉGȡd+jAkCA5FnF~\v[/B i`Kn7f#oi}oLꟋ;GlKڑ#%sk2OۖA A3SBPĮ1`/dP# 4D뒥h);YBd}򅼭TwFnFxusAn㥨m;=ޚ3MQTW5% k( ȓ;)e PPXnI"da$NX`M.J"G(FJL$..Nafa<\wF 炈KuFD7q O1%sNƶUڳ/mT5b%]gp튈R֑j_ّč31$ U0lΈ͜#Y?ꮑœ:ߛLQ{Z DdmӦ ȋwSl6a݆cBq I bZV#I m8`MiYFVgcl@dh 8r XG,z X-$T) ` "UjA2Q;) T|RJA)%ZB@,ov v[QL g5 ZHZ(.\-ˆd/``rgꦜ4<~*7axLy]|0^}rY^}O/ Dd~p͘,YIn()&8㒉vpd\ quזZ͒⬌.pSI'QiVdzC49W~0NgS:^\.5ׇ:iI.Նph7=3Izi%fJ3]=X^܌]ŕ ߝ?^|v@S®9y4>G! #~Y{G0Ho\ e<"F{ Srå JriW^LJ4Mgx0ʎyF;BL{`0mZ^ wA6Żxr6)-8fߩv+qӳ ;-6M+[+Zse-hkmlz"E?B.HP) dTQCEH;gth*9^52@W<@YO`7 ਙgg6l R=RVE{ݏ\%w?jx^rҨ%'KN^r;zbOe?ax2/{#Rih>VNwZفg >e;hd\E|j2dcA&RR0e'Ϲ&İIՖ+| xО%[5@*KΜG|,+ʱuLBE~ц,*T#Zl|[gawK<Ì㳫9?rx7f=fcJEr%; tvC'Ё'h܍}+'ľݑoX A*U*4xbBR(f֚@_-musjL01Eҵ/~MR& 1' &( L[;9[t̜cܧ0S%aHej*Ma ʦx PTNmaƐ5o5dje$S q^獎In5._79;][2Du 2Q)$y҆2U3ϲFHѧzEݰҦ qLjDfMɐ$wU#gK9ۿit[.Km wYV2;l/Xr)w$࣡G4)_ _(-?F dhXLFMG' d0dҗ+ jUZ85;Ijk e䨂' Q$KyXh2a9)Dkث%xT, 640 / " h[fI;ڶ)-a!d%B,B)ˬ!J4ыUc]OUU-'5XVv[J,%Q$;ea!x*&cƺ}j[Cv]Y:=hsWGo_@HГe@B 6y`΀IU=$g )R6꫺} a'X :r ;KСTTzZ\m q}U#ս XhrXZf5.ʁgFGB)11Q 锥2` 9rL_Hɳ8٬+-_Ct>^֋\/GO1I3]b.Z^tD0W*-fG&uLI:k"V5%4xDiOEK}x,SC<2)j%kaJBheǨ- 4Y\9ŔU`K)2D!8,Fgr @>w*R^V#gˆ x>=klp;+k=T̖oQ^bUv؄<w;_lw } o?dޏIm?#۝n \0C/wwosfB;}rk>op[t!/NwߏjV>3i<EGw.)l\'H#]=7ݾ;#;]o0J,7<,;K,Q4&^2 Jdgt«j셋Z'!̃*z"xR5vJ̕jlf@ߵWA<Xvv8]M }YkԳLS0*8iLU ;c:$ނ 0}^ygh쀦۝ρ .Oj1KY$??k||)߿j'x5X߾yR`QfgB NSbY{TAx , \lBZپǤ*e&cRZSl -%ԶږPj[BmKOZ1lʾ)ܲCt.qYΰΘN)Ǭա4 P]mѵj}ove-Q띉xjtd!((H$i PS hiC`P['2ȬѰ#"N,j' 8Zavz 'Ԫp>jvRVxx鹻wg)5(te9(kG\MB1O!z]SU,OOw5ص7T-Fl, Z€x" h[fI;ڶBV"$"`4r/Gh;&sUc]OU>u!|[GLk.h 3W[J,%Q$;ea!x*&cViZ 8dWЕCNYXя6ga-]*ǞK7ƞ@D,Zp5lrK9BCGvӜ`-bmkwm$G2VTXr8\\nAwW5MHʲrdHQMepg駪bgvHJ( x`HQqO@4-輲d^њnr}ݻn˾W߽N>NWr?퓸{䍃+ )w_yҢMw[3ږ]QGus; 5t/ԱgxsxZ7ݝlaxǶvXBTO)\uKK2]ژ - ܟ6L}A] *fA6: B[ (eL9Q ϵWu 2y%j Ғk@eH6NЙ /iAV&nl*;BMgǂɳcs1~ncc|v6I/fO,~\VvyPNokD2b3!(yt2*T&QA r)+EA"ǖ:÷ N FjmJSDTLNGIJ?ωdz2r=t23*Uԗ<.R3xĐe 7AXaWB+!/5xt8zzg@7g iQ+/G$,Q$MI,>#Csl,a<8.rzg!6䅑k}Ǵ#QFI'*zQHjIK2(AAic}PeM-ztvWQ82mR"E-}cO_E'ݟߵ{0R`*cp*gPk1g +`L1XX\F௪wHujr;okA6Er/E3٧4d_t].uVԍxovmԭ7my7;K1r9exndNLtCLN.>2E;B(#ʕO>–=br릡 3b}g~.ƈ;F&mYP/|L˖WsG 8O~3Ic]Ɛ1~*L}]W7վ;׆/4/ۿU=e a/˻bY c*6ƺWIQ]t7ח!P^]e[2 w]#wC2wm+!?xv b9j7_ %t;GSݻg.J˿\%sZuy0s;g\4}vӕ\?/Oi{UC{9.X`uމɲx1z>ݫKJL0i#P:఍V|U=jWY?O/z^ec*£Ǫc*tyP*_)/2L '3N~( NB4^%8omTMРd$Ck(PUAw[YBʉϜRR=7lUFss^}WekamV]vR){y֥BH?8ݾ2.w߽zϗS{WP(O&l') >,=m4[Q7UR( -i!Nci|ε@{UFKA Ke$EZ7̨UE[j5EWl2 0ßu8%%ܻ2{'G$l=Rpy6Z}_g?Њ@+ڞ#3#~lhi|A'b C,62k4t0!{^}ϊ }̞ }>gؽjunn,ï0|P3|ئJ?,fsF:&$ g4ZUrH ], ^L u_/˪p#hVZuGU2ճ(#/EsIuV@nB| gБpzl2䒰A&JY 5Db9/%M"ZmY[I*ŅG|B ee:C&\VQwژEF&sj~p@A-=xXO.<@AG/qp"8P>7铏g ,nНѧOI+ +,\,;$/7JXp1vrGɥ{gl-q.ٔC]&gCǤsIB&>B)ۏ6埆>>o$>xbA״ak*'t;4b/gnd^קz_eMZi&!1 IɤSrL*qV aP=%놘GU6x\xp$xzC!(lN:C*jIrdWv g=M_~sᓃ9O"4,fmDPv$\b^rkJmJ{ʝ}}lVE5#Y0Z6am0e$"Zm-HKu\$UQ[K'<-䂠'E2]d})ZzxRXMg;2*հd싅2€/.q,̸Ywu*_v~=]~wB|x1R?H!rD7ץ.`&2Q'\m:5WE 0 "V}TFD3  I<0+cT1*踳.^ϳ'dYtC)̒ "R@r5hF™!Z Ss#%^ސ*2"VvD:K||դd_\tq8$dK8LA#r(#RNJ(71CAr\]x4Zq,xg@ؖg]v |9  я/Q=qE?S!(rKʖ,:B6AY< .$~`GH3OptWA_=yn A< \Ee2IHNFU6q@Ɓe$X^7ZB)q!XmEd.B JPh@kj1Zs-GI2jt8)39 -RL})EYo퍧w!ԥbkuEI)PHcɴ6Jp@%p6I:F.0aAx( Q|C&/%2zo0%(#+L۝r5҈IhK@I꪿EvgxP[ʺcFg bZV$ -8`ɚV&uϯ'qZvQ EKb#Lf &*FtJ(j8R Huik4R$'Ty1(eV@HB]]lFӰ&??Oqa9~Q<cjFB\iZWm,~>rRQf>kH{ʤNG/I5/]=d q ]ko#7+|m+|L6 $XIg-Zllْdl)."YUgJrN?qqA,LQ=ԒD YG'$:;P(kՂ w8 ]QQZ,>4!r:ujpd>-eA+Aas3Q)lrV%W_V\5{*Uj?:TLmFXn:}CErD# sct׽-7cQvR}e[ %Aj[fX}3J6,q"#G1~n667ɪi^*#[uպjŽ&D͑ܰ@~01WszF;!J]?GZ|]?||nvU{,A[ tpi*+JDM}n+$Çwo?\2}_/>^&u"L %¯{ r4?ogެiae|v4vԮ! ߄B4r݈@ݾ|y1y+F'f<w1xp9KːFKC[ %Q*;k! whR%r?TPzwܻICETۖ95($#> jv:hn7;zxXm.hu.lȮ;G~i~\{Q%2wԊP r8Y\N%ycOAMA WHIOP` )2q*p=~Rn++ Z35nozX|fo9dx5ԠW4~-ޢ3tfNjRN(+)Bv?UgSJ_h04߽(A>/ِHK>K~0!o; $Rp]ؐ F)x"Cy&,qzq*~VSUhap-/RY*MMm"EȡH!)pqJd[T{Ɛh ;&Of&K -,֢ -ǜ~1 ,?ғـ\,%^ \ΈVTfD0\m'.5/ WI_]m' lWծf UxAW \eicepPMz )Y>tv4jKXWo=;コS~,wak¨~jOu؃!X߼?ߖ^B؛MtwMQ]rŧ$G MdW3[HJߢM&$r9Ǚ~!`7m3N5 r!HD3b 1Pc $ƉLL7I$.}HHh@fR &r4'O%zְfcwDL )\ wHi7#>(fƛ--sp򲚼O kdٗEoո̓|Y׭`Vr63JE9LLl֗g:ֈ7lg QxUö6X[BTOsp4c_F H| GlG %{Ǘ'ƠS O6:!T3-<3Y2-B$RH>cgHDlL93ZXW|?bu5[];jiQ5<' <8a 3FyeL[Bpb9Ɂƒ9l([ixUhzZyu42H[K> 84 >ACR h .[̵w1i!Ҡ*㉕QF-Hې<}y ig}cѤ/'CjQ3g7ML )jڲb<[_j =みWz0=,0!@{{gDJ6))IcVNB>PftT՟G?LsSGdP{|g?N n:׺!4{pQ Kwei4S`g*:*,Je}Cdxp_Ç/{Kn']v'Wn5͘O܏¹- \nak>Ds >ƪ*1NCM,;}e+N .~6Yu*ۡUR'۫iNc܆ؤim0~+fW{zT7l ?[·(~.sh?hlkyջwWw12RǹX"KsX˹W8۩OP${鄨W]uukˣj/z8J!c9~zgy1m1 wb",4-8xoBsHU'21ҥPp *nڱX·eә GOT[nV,w \efgk*^arTef1㩚qYE6Q{nEC3jbkivb`B4Ci_{!P3$9ud$f4^Jp)3Mp8hdCz|d7oySG_L=ھ>ԶM~GIRB;72ټE&`=灃򌐠U+ct=ktYv'zneܘn<2&o/: =[mNݣ4 WvQx-oc+ȏ=z/ql ק2ҜLPkWjy*AQZM>?JiXk j-,{:?C!˴K_q9I DLʊ c.(\Զphp ~I]ln>:X℥W`u4q,f АhhqHB, F(5F#ur)\ІHo$с>i@30Dfv  #@:ݸ=L(5/ٚMrZgڙ-mbM #5Wck̽1'2E ]{!F$\u q'ŗ`o!R&s F"G {}6Tˌ^UtЇ4ɇm_'`U{BQ^ottA;b2bZ/ƃۑH%HPLF^Q&("+P%Z3+a`EK^R᠋{,Uki#=kyԓ5sxTRRTB+N qz}qBdY0yY _'=q`Zuyylyܥiеb"MΦ>#62:%N\8scc IT&H3⩣FksNLIExjric?{BvrU%ywk:FBҠbiL'Qh GFS=(y@O6{g* hښ"uI{S4H^r*yb~/Y`Iydᝃٙv櫱)Jr]Z/cYJB7#涇ϧ? rDB YʈIKܢU9YA]kP~ fP z<N4oP2b!zF6JDb6 6HUu ЬR MrLZbbȏuMMruPHY"!hNgrEn@ j6&ΆJ>"nZM#pA$jgQ\ ~HX0x)DeY_? /8MU  2O}D1.@)Eoׯ)qOXF-0ӣ.юh4'?ݕ7–CIr>DU̥UYTT&rx(*QSxNedJ iPDQppJjKt%:" "*>iaq!qz^ q S-,NW426&zdl4iƮX(Bh!'D1>+ 0m(dg{`XYZr:}H-9䒚&:iZarkrT\ Gl'#c))dP82dOB72-C1\Lz/() hd"[f69̂ܚ"~~\01wMQ[7ںEV]Izɝd\r! ۊH܁9!2)0f0+mrVB&͐YA&CX# 2 G(ɦ dT'㚾n6I 0 "6"iM-"x`(CP6Ơ/ςm*[f)[CgGz26JdL`U-rƴmd:Fd7YTptDQ #bc܏'^'\NYgc\).چq[\lqqVb M\Gc2 PF {QxgB6QنoZ\Μw6v3<|ٸ[G{YQwk`!7 {Sܠؾ7( ZqgN^vt~tc5m})|X٤b: %'"y0+D"!MzlE<V!$a3rY"hcP[IBrOQtL"7\δRPd񺔸⣅z)U1 e0>s$ӊYӹ|;HNJ9'Bwyrd]\aU`_\O꣮Lk~Eȓ >vn~X#, θd~2ZJ.áө]Ggɳ鿨Yr@ƒwD)\TbFnIE?*s?Vׁ)v\.닅sj źl^NϮj$ ] Zѐ5I3y'xJ՜cڜ&;V>u?~bt2rv1V svt2^[ikpPDbTg {k`lH;GuHˆˇ1z01(Ze'{WӅvG&r騂u:e.+TsN+42"a}W,OO^Y6{Oד]xğ |Nju"g'-/hK+j]ϿƭWwO=]z5eE-Y{vy'1kr5?V>]}A<:DdnyۼVWO&(맓RU+m@EQQkQy/ubcoT)tV c&Dr1βuoSDIL\: )ȁ /ٔu 9=O D8&3B謏j|`8m`)jNY9\~ef"!-cj,X ^VMmBM FiĴ諆I)}>ɏ WQZP5R2DMb^NWbWԷ i& Yd+fL`|"ǔ|@4~DY)#>7INikLR&z˘ˆs]H$m)r(9 qb'L#HtsZWDEޤ)z\RlrL(0&Ef5C,?$ZDc R㕷[mLr+jλa$#8l@?FzJ'm #ΩȭG' _֨94B:(ҧaLDV 3dH}Yٰ5FΚr{A]` Fǻ6 J#Xe%X )$DG"Ф|y)?5p 2`XLFMG=' ,j+ XH &t)HW*VcVZ;nDjlRj,ȃ2@H&Hq,br蒏l N7/$ M(h }Ʃ |0em[R|V'"d=lDh+YaRKqE5" jXfInH\(Q2/UI+)( "Պ#6dAeߊGs1Sƪ]l*a!Űl5Pn2gxrjExLZgr6!Q9B"XE9BޛrZ^PiU[,GJ{WD0*>iuڭP؛pET*[ʒ&`l \ju*T*Ջ+W+Dϭ]='PWè5yaTiWpZsn֞,*J/pU;WJcZJreq6 ZdpUo$aOpUvoઐ{c w^*T l+F-umMQgʘ3u%= >(i/N^[KaoBcXg=MM\fX0+UI.0}4yM#~Hl1Y\Rɲ-'+d'O/:~Ug.S?:/?>Xg;B!W BL%"ZLQp {oL7q{ͷY%nt+{llc] >iZo~8} ?F@ֶ+߭o0Rxz:|I3o0@J-**gbS@wJ!̍a|QbOT+@Lw\~Xkwĵ}<]vG'nxR.|_Qkz5AJ)z)O ݿm->w2ny&rI?P9)-4jS0uC}n\y⓫O7Fg@**$*,V("Iojޑkݼ,MTk̛ݚsY_kw:Q7rOiGSxTW,sLdWY&"L rĸYd@h׵Q32uspe".Cpi 6?I g\bm)4P$;#jW/8t4-rs7?/ (@|vGm2J#&xs9 a+]׎%l+7R)& zKrՐ8lt+* iO[&yBVmM{mXnwoLTSNh;k@7Dq؀(p!9QR˸Wmw[{W}+3"cm”}1X_2_E0LZTKԀ4>EZ€ʐ2e t感 P&mXCl5}{*^,}tv2!VlIw!Vbt2nP)Uϯzk$N:(ѤHb|`Rpd)\++J(Ugz㺑_i>,o!lyd0bd6RnYQ[<}QK[R;R6Y*~U,VXRlEl)D7KiXyȐN_b ɨBUh̭9%x#zr\}mIV^| wlSkOo+y,K Jj޲W_47hyga?Ai!HtPT`2+1CgCq0(墴P S #9FBYg Id]Z)%(׶52^pE'tREJ!F/"-D ^59Рܩ_AC_|HE }0wd@&DU0 cO( eLm+go4,ŧ̦⟵?Qdʵ!ZP]&j  j3rH?bX/kpysRQWkTphg7}떔ܬZセg̓>fkrn~ʽ L _iSߣfHJ@Mk RVXCwhmEo8d8[J$%-or`5è Zkf׌J3]،3Յ.A>.\QT6xK5ޝf&ޠ klI9IGb]2O :)b(1-h*P@ TMel2 Ol"{%/>7%"uZ_cɘLS͸Xmkm3hx/E%ҁ2E2s`s =  rE8ma $P"[Ie˺` 2K#( D*lS&6 τ_Fs<\hfqF54wq< RҞ(i+w>׭ވ\K`[td'I%V"Y6"ۍg"-L#ʎ @eQFlF~x1󫎬BNٌKՋ^t^NVj)\?WF TP[""Z J: zqOي;NE#U}Mo|{/%AaApC3E?"G~\F|<)C5(Ig-:]@ֺ2Îd'x$}M}\B%,8GX-aEkz 4: I;m%!ZdtY" ɶњĤ2,)p cpEt6A{F&l3m s`O&2yTo-P,u3mRxnmg㹾Cua1\[{Q&=MBPbleM) 5 ]ڗ! Aί5B]YYEJ`kpժ$,&[8:yg,9 dth9 V{bs ֓Bšh[2kL19zc1 ! S55O5"H3H@&餕*mҒv) ˔R  Ys2R[A:1EVWJH5%K9du})j킒ȈYcRf+h}u݂o$Ͽ\ƀy W#G[s X="l4_/3kVN0L?ꃙ5_\OzSa|AM)?51tl T)xJot/͛錕ś|ų9 sVI>ʐ1zD$[u=t rx6a~Z_'2h{O۟Sbxem_:e=KyoAxeR./WW|5oԼitgN֣yf_{wO|{5:8@*^=1׳q9?q߮횣YmY50^-y6[GByma::HyuZlt>8! B jt)9$rSQOV/4l.@M`:@IL 2X#D@6-}fr/t跞maўΖyvy?B(T) [EPeBJx#}eBRJ ԩ͝Sxf.J~$]JS0I80Ad2Νӊs ꙬS^Ͻǻh~BۖffYk) Z$c=9x]1^}&[jWYJ>5JKSHyeۡsNk/3 ii6gN-- #i]c4@Ȯ^֕Y$P_`FE^Fhq"VI7A4%>fBF/.cD[mE5fN)6YY\ۄ,o/_<ƋIrrK5)~ tL"KOVѥ4[~z'a>XCff-"U^rw9XRjU:<+PBGs*Y%( d1giBವ^(Pr,1+j,g9PNr{؅ʏO:V`t{C ^X7(}JB+Gb ; "R>_(3?FU1Ɋb"#m$ Jx@FjJ0JgBM>@fP+;iݎ<;ѡ(eNDQ$2tVL0p)mBSIL.HA8(݌'^ɝ`M ^˜dTĀj>CS`vm,b&L24/Gi4s/di7ccO7Ur m]rNŒslՒ JRzƧb &h'+0֝R1frta~NЫiXS8^7t}nY>D8Y$ &lnlB<"Jy7_7;I0\bŧ.] &mln}TrSnhi]":FG0lY k@1kQA 䢴65JkY|$ֱ- !;R-kPCf쇶;Ѩ4]3 'l}||  9Ĝ[}j|^ Uq= XIty!cKcK< xIaQ/$R&9Bpd Klѥ45-] 9SD"YBDޖ&Cu4501lF΁)_.o[kW\Y(>.^>ڪ}/=]/vû|}nu{åP"w6{l)T n࿩T=-u?ں=/xUq;W[vܲ[pZ/zmEZz^k.'-vag_1|1/&Z>mF~oi*+ psdU{8<-rg+})nef"@bDcOQgR&mH0Í/oC.`sn,URL IYV~Uϐ)sٔ({8awMwUSOIbH"D]GMV[x9ٮ>ʋHR gsSQWU Kz!ʼk( a% pPI <XC#w&@h)JFWnf f|n7x5D3\Unyů>O}wC7x.&%&8'p/(I+OX&/T E{pu{-9@ A;2k(Fh3NA)#E,8F bfZt=v')4C$/;\Z'di(_Xt<@gZdfH*͔ƒ9h({SLлV{.F)xꥏ<Z&!'Q* ġ+\Ydn(ţ#ƗvAp&.hq _szOLZg!`4xbeQK!6$ϩGOhzWYzo-ƺ -Zl+(2'ԓ@QpyR ' xv:+wsX e0.e@LvV r "SrJmۼwOwuG&hϓBxH !%ip 9i!Q(q;]7Hm ~|RZ:y>>ӘFGDַjLfzQpB(_:.&ëqLt6D-/>7r_[cv/}RVؽ3ؑNLOWʢ9tmm\0HIweF ).ƦF.i¬B9A9@P<*9EP"bv B2Ò7ܺfCb*e\lg_+W=g,]l=v;>~t^:а{ehnQyg')a~yb;뺽q¤'~ӥL>-Ń'iAKlNpW3.Z̾}֨ɂ5|}GeO֝kY op@ng7 .OSgKA2/E,Si˕R2k%g&!9ysٙ%t/PVC̬ow-{IRRxl5AqPoǙ,!oqāѽLƱ+ОN(cl` 32q/V l⑨!N36.ͩ|n 7xZymAwWûA>-,Nev3͚5=VO?.ۛߙn:o0m i'~v\?`(o3ݧӦ>mThT_MLv lwG?cj}J;S;6OÛ[f}H>oQ YMS3Kرx;u'XFM[k>ŬsW=n ܀0  BcSDj+ VU@Q糖s:=5'ϋjOc\Q(K;c9cT AypD͙DD4<lNDQ3k"/W %oqʂ^ הn%sZ^rnd1]쨀?2W^/7UC]r2 sV3DjBO^PFp\9% hʢdܳ+f q' #gXRܱDX£=IEY֊@ӖA<R,I+ƙ).FؙT|Rp=KBIǨs>kxa;+&NG;s2g-skv(j R$K 8MQ:kwyyI^ xOb!8<7&k_X10$IxTE2!AI*p'=HA-?qR'S/m9 oI

Fesryz2n.jE|f6g^D_1Q54^1FQս//^~d^#/oc&{xF'tٌaȒ͕3r NT|H8X_pZt477;Xdo]c6JL"Z526*_v)xRظqb+cͼ\ۛn*̝xF7wd6fr_OSNFU.7ϞP6Ww2f?]ޚ 6_91xAf;pՍQ8zɗ+:vy} 'ɢr~SP>v^S\9PmxM@ > ,0(c4 (|I'@ą2=msa=76 $2AB+<#:j>1aJ*Sib?B]mxs";])'ຝU9fznf;5wάM{gP볙-?yc\׷#xUɱ KOٸ+ Rim8.PzK3ة%C1YkI.p1HJAM AjϼJ"wHTHL451sD /qQ5dQ^ E r<5zZfvzʳ]r!-gy >P&mfXf4X7=m{+.<rhMa,]uc3kkǧvXeOWrt+:_ԃ;/Z~xDeekOևdx,OƩ[é\= '㵢 '.ͳ;rW|]s۩%)ç;/C#>6\SJ_rXWã(=rCϣ-(LQ|+1ֺBx\{[n>_\S6,bXO9δ TFi==Esn{؉Kc %@31@?1Z#ͤL <%OMTx*he&PV;)፧:c\k\y+`$V̞b1q).' r~f#.w Sxq:9d[U-s~ַ=<|m>!%(xDD NH-} ĐDщ `퍷.9ٮ>*mꍁ k igkk6Yo^(ᜁ9&Q:z1FbrpQ,^Z}pc% pPI <XC#w&@h)!&J;?5li[;UT=# TWI[ i-\BI8BuP:n̅u\*:.<ʉ!ZZqUoz;I xf۶ .`J/^)0ӏ\}kR8B9eVwDٹ@3 ZVsP>EᓍN.L=@LjpI(o"%$aJ -zB~ '&gjS:wb| n)j6orj{ֱr?pbfD2PzýD$|a@PAsFrnG!A!r0t@(3h@rQi&4q#OA)j 3AhTDӍ\pFi{"ȼ-ss-g;X). / <(0F&ŭV!sQ{E $ifJPyH@m(;)<"t =|u42dꥏ<J&OTHmC(\w duagO Ap$O^֙h" r0XeR$ sQ=Th|0]clzFț@{A+bvu5GUEj`WǦ>@nVa?x9d-}P*ՇئV93hkL{Zm3[<~07R61;7{=YäY«*+yeqZoimy;Znkݒj$ ֮۵'|>-85A.Q m4o=kesMnlgFgXLwomKHV*5RN}y)Glu7bPm (T*K.!=fU`4HP 0}SBݺfCl_е2.VCJ>c֥퇂z8M躤mh~+{ȼ4/21S/LB\ <-[#3dQVCLRSZw^e*EAEcNNx3Ok6ignYwμS DpINa.Y"(ɼ5LE ,WKɬi )E9GЩBQ "9|Jl/›$W))80f(N[3BVi .te@/sOi]_=y[=o+ۺ|룗WUn||,£Ǫc..qؠpDΤ4e6;`/_XzO}vX<ꬄkfD$Yvmtcd}e $f tW£w.hO8(ڠS +Iʂ&A9&-em{D;M HŃ )C[!ho[#ڑJ1rs |#O E "$%CE "I`Yƥ@ShUQA@}_:*QQ$wAo!Z橏PHk T:RD‚Z͂~u>`vK\[ };o+GlC T#xB,*d?Y ўB)FA+\ƞ pg6T^3MdÉ6`d1ta.FfvQX21qǾ- P`FQЎS ՌKE9%5eۂP܁X > 7`aRU^)!= &DddKTBBңQ92a1Ǿ #CuDs 2o&^ϓE=HL&pyHhG$t nDqy haJxԨeUU5p9[Ҿqq:?&_g1.Ma\.v֣ `b6tʎ:崓!PbCȜU>) Z§@.6x4Rq,xh@؆[lͼj |GaUtAp]j"]AY̳:*@*z/~~<ԉ_ /}8_ԯz6 "_1^'# O9zG0_p~QT Ka=J}8C(#eU=bu%g:pE<"^ _ \AVCECml6\~I:yOG|]&T|~3lb/ƑRDa -䅊SǴ.Wztir"B32FJ8Nj"0hK)\[Q-|~4ȣb+KJ@dOQ%fxE6Byb^m|eYhq]4U#\Z +M`1]619eց@si`o"* vAh3 xp%.g*)\F)߁V oTDD%YЏ]b(df n!&()30I @En8U./HrJ? OoG/5mU"Hc9ـz>䲻NyA[@3В Pcyb+֠}Q;!TKP#< Yy{cP~~sBdWG{z0 Mq>5-G[]_54 !TC_^q'gwT9>yWOE]s]Ͳr#b1( rs7K"dяGo.h18z$HkO Vt kFn= bqQ8uV~ƣ|w}.u앑kX⎓4e$w,}O|nzz޸>S]Vj/j= >̫#KV(3(+:s1L{ib"UؚJ_!F\]闿?{wo__ޝ~2}o޿{'\Y植:qHQ|uo~ܾ]c󮹆-EGL*(9~Я! _ B4)ŲZ%dkz!p CG)h꙳DK h<-JQ*-E-YIY!wҗ8ѥ_ȓ2i@ yrc=&x lA2*XNgZ' >XS`w-Ѿ;G{8Un|0BB.&0E7Ӈvվn|_Y*ĕ պ@DEmN)O6?!z\i,ytxȤ>F No88Bc 0*IPMF\kJ=!&jijygLD)F_(H[A<J{sTIyNnD)~.aS @uc4#=~#W2[D0ӲgZQ+hƔܯ*"(r{wׇpW ] y8cb9'BAC7Hkh|3jdX'fN];pF {0S=в[V}jգW=jQ-[iR=J%G{Yfiq^Ñ(@da>Xκgzʑ_4cR臞L/a Lc/ ^ڲtܿ~G-ۺYl)9A:4YX7E2td\@D<^;ϓ!] 5R0 OjB > faQD-W>x xОϓJrG<& %2p!.(X;km"Cц\>jl=E vJ'>~{  ?sxMS}HvfVAzyx=6;N*A`^46$(aExP]Zz.cOVx5i9ΡrL̉x>v3cVqVL^V_zs!ꨄV'ĘȤȤSrD(UB$2lFU6xq _Cis~F\hU5]}]o)Cl.dXM3%uMdr`1 .sJ83_Q.\`.4'ۢR)j߽U~zn+N3!~r>^_u)݌rqVjͶX_x6f_㡿0?}|,;q7χl_27VJ%^?rܟ2ߡBfyh,=q RFs& }N.R{`lR$j8 $P楸bbقƝg>6N;d:)@: hH<"`.#2 Y&pv v5r ǣ/Gz--_VAS8&1*؈!ey_v"T["ZI! VGL<&̱̬fWۋw#$vlRc~ɖԸ ^L#-m690` 7$M[1pQ34VA"] Vy @d *& %B$NuwS!9c|q:S:5,F"Fkt ;2 Is&(djtSWN]X_vwsb(C:!#'A (뤾/k~>V-G š(CۈzRT۲Rò/ Tċ=4QwzWyOxWs#帾f$8)|igYoyq:jSOߦ0ã}g6ǣޕm%*JfWs1:h|9f2z9Ѡ*܃Bɍ*Aȯ |EcUJ[Ef^FÇ_C&L.pG t$b\<-ufοs1&KïaB'%qlal{ /ө% sDzNߐ?.k9ntO3{%}TY3Ui˧K"I_iH·vU^/#ܴ#2¸]BVLYsa5?Rh`l8MʲazRS\7\xL6jm[BɁi:2plhnj.;5Nf` LRڞoXy} 3jvyeN.9u+u8ߵ>d[~}#ē6õ]n =dkϟ媨=Pa$'5 u ZfH+hVF@[t-Z>.PȨ)P)UMYrAЃFL.1ٜLk9WI\cmdFvdTj/X<>bz[45n6ȻXnş5h`0cПLs _H!JeI!\3O`ɞ")&@30TElieZb$ AHQ`Sj5 DvY8s̢9;i<ԮںCNݘjɽd c\o+F8 q9Ĭ (2ì1XŁ,d 9%š,H6jR R&2QxXx,ǂc_D4t!&feʥYtYe 3yYF]%)̒ [x1mlcT ˚LdF™B@F%U#i̋PÍsH:.N%⢫7zI.Lgsqԙ`NK {Qob6ɭr@.qhXpϫa[Ez{VoI9QlNzQɎ;$uُouzB4&oMz8T4e=rn}R{v){w@H| !OU2#1c}V%&-xlMhʳUTV; 9*Br2rF!1, ٚL(THȅ` -`]୷AJ:P4MqdZ[FΎٚ|~4ģj#=K %J+?ݎ# ^/ozQuq,x\Zoir2NI.KBKUQb(! Eg$cdJHJB/:}Ye'L9W5R(LIn1JNz9f,:PYù!@OZ"d'o? xĽT["ƽ!n=ÁG=ZȤy`nݶXψ~<\dyzݸ,W ̿FP)?h6p2T#:\^5RF_#jT/pT2j=j4 S_ .Վc'j?;Z5~M;h @E4IbE"Kf_m ?@B8LY9OXX ENr5ی6 MQvTd2>;NJE5 ;iV;5v(G-$$$Ɩ+ol1eZ1wAl"&$R;c:OZ#QӮtR!%g%VTsɡs'> [,x%eiy~Jł7jűea<;c.e({G faGւ˓&!7yyO~|+r¿N`#v.(sƛ~J\&|V1kJJPpD$f5 H]YՌ8_5:y`G` >Cоh=BGMg t[gfs]paϻ" =s ~s0DdƦLˆ1F J'6O9W-ۆȩ۳I/}uBE|1JHt/{m4Z:L-:nğOV$NN/CP}j2I=$洪f \ςeCߌ͇ew]rec]v_`^q\tC'X:jm 9ZhgŶ%{"H5Br9 ם&(CF8%8ZS+܊t=U $0Pc[XMU5sd8jRB=A*&Dfe ہ~};]d7Q;䬣'ǎ@jd% E\nLΦ>6R&:I.~ 9zfl 8 IS#FkuN_,.]L3p?XJ+_ @3?g@:}{r{8XgOGEKP[ VԜR#Gn: )נ_܎xˮw,YR~jڌּzŠT I;%<D) "%IS>(Y'˺{CDrY#!DAldhښ"qI{SD)ĴQ%K*&z ^{aZ`rhk!ǜ1h/-d3K^bW0:Ml&HF} ߧa8(4sGu$"9{M|jvҳmnb MoP# yui~3 `ӯui3xINԁ5z`2]t햜5JUs?єн.iuz2}:GJ ]>y<*H,* Hk->e>'HYz uӈEN~ ɝ4cZ9L;s5pV6 rKҥI3`4rs rIW [,HYOzH9Tx\QEmhe\z<[|(~U>E')aJŀ#1!qO^J0X9% hBd1|~rrʻs4a`OR8 3kE iKy$,\,N)hH4; ĩ:*e0$4:ؿ [ Y1q6KvKBQ򡤇HLp1Xg .//k#9Gcǂh g&Pn& =?Y^$1 H 49,HCnh哉j=Dv A-?1"<ʙ!߻r(QLCkuoqfVo;gFMШsP6vHhf܁IV֩p] S!GGeD016iO %K0\ 5IXeI$aJ@)05 R NĄ2ʷEaF333 d+rPqwj8'nR =-T'M&m~˽I֟ťXgieJ,f~{~(wF2vΛnn4ASw>?\) ,T82[GmdipçT)ũBpSt7W`-&@ܑ8b!hlLhwmBm f >;v і8 1Dࠡ;NK.l]ZqLs4&Qe',z2 sbwX隘yljSadrC[xz꩷wtpOҙ<5 %68amtGJc.XB+e[-bC]Tk'&OMO,n[bgvvmOo s?0ac7ڔLy17Y1[=azwAƒI?BM->ϛbٻO[ ]^8`*qZU߆N>|޲oCۻ.)SecqCmv'IOwFy1u4sIsD\UYG6 7X["Qb2}4YKb1qYYŚ;4׃(5, sNB F՜9l 7ØrE&s![#?>(FT= 4#+dsؑQ2`{6.]B?%ÎvqU )JֈՔ(g YawuOկ$BJ0KLA'&#E鐒7v֖17fȏsw;__/X.D焢%"in2˵ 3o9ꁡ@ S E 94B& g`S1FLU"q7c\#~;y~f-rONssJDm-# $p(IqU.qB S@•VíV@23B=#Wg I#CNz#:g4(R@[2ן(M&x ,113TLBkZdž27ė.b:ԢtIj/ewLVɷvj ob^8g 5ݳ<[V@_6qo)FY33%<gfWJ#{XOSgKA2/E,Si˕R2k%g&ਟsJ{KG"Th[x;ov<1!" UJ 'fd8$- z!ASD7Q@H29Чj狋_~_Wv1޸&[3}qN9q 3 6߹ە?OD8 + x诒?Vbc% bJJP\XdZ2J@5 "V'] /vSm|ssHkM'6DB-ME Θ  pN.>.2z-KiR=J%7m;/fi8+@dY1yY*𻓞80*l!:z{+]RJx$ lJYgD#FTXg3@$=9Åsa=76х$2AB\O5Zks 3%ᩴ9d7Q萶 v-u΁%Q^g'x ` ݾ>R4Tsphͯ4۬!; A9M#C.J0 B`ʹ.^)Uh)!zÓրZ1Q8R 7N-1&F*Rt2}6W ǜ&>9kwCe6[6!dU(`RHT (y&5EJ)fLm?ɊJ7T$52J%o5ELhR /EJ\|t2x{VtQ 6*+z,KC,_KlYQQ4vT7g4c% YʈIKܢU9YA]oP1 JA͠p ʝh)oP2b!zFmlEmPU)P@JI*tT4 1i.&PhQ?5NS4]]9C"e`T9oM+rkZHPau9[wB馭`@hiP BB^2j"?E 00h r**2KG%J@2>wAo!Z橏PHk T:RD‚Z͂OP>?>4;...&q&7[#gK8yZ7{'@@E2HcZ;)Q{T]fB %%:pg(J!+:$NBOkB2tJ`UR*XXNW)8cW,z,{ƏU"kDZ;m̷&O~@eGp6PU^+`8 ~Q82Kc4S(Ũ hYb\te4OӐ=12lr'!(fȆmc"^qnٌn2k&桠v1eaԖ=j)A(hǩj%e2mA\(@,fћDt0e0 wZP@ 5C hE{bM-ȴ8"FBBңQ-'\xXlǮ #GĻ8 e'PMK$1דE=HL&pyiHhG$t nDqy haJxԨeU(y(ٌ"j8[;$_g1.Ma\=.xG+N=څlui'C<2d{9|R N0wz\lq`P솇{ [,sdFD=n&1wӫ6I̓&*@*/~~2 O6yI^_!#^B"T?a w<~{2XLU?|0j/BRqrQS& G#.zZju9xW#^l _  u#ՐwE[C"޻MBGv?,x$ˤ@x*T56hw5GW3h,SZ48/Y/4ꋩ $>:[&_-Gd1]z1|B\ZU_^=C-\~h6C/P*·~ rB[@%FvGE0#woH{@BEXf>qarGH% /4*2N:%@:RIqдlp9s@H#90NpgՎseL4 `.(FΖ+>?@rQa5% S2.8M渮Fzwټ ocZvɭҚM`\iܤ=P1Q( #L|S)TI5^ o fVWeLX<%xrw,$m@x(Ѣ DErsF X32iF=· ۰H (4 b$ J%NrJ"gcsc-MRXNF6j4()=-Y@ :zA*+H9pʕEL^ιr!g>(/;!TKP#y~b!rGQLJu5 GqAb8 )i="q>|}U1i*ĩu2Z/f6KNJ( yn#0aLB8K' ; GJrN?Y7!~A;Gt0iTomtmk5WS7O_Ù! z /^a üe&AX~dԇnyݛ^#ve}.E˃}k]s [t--UPrF0jjѨCN+횾_ 8d_ă4Y%Xj4AR2J峥; 4)Kyu?dPzl]Z|}^OFb xZȤ!"*vmˍv&㩃(tAɈ`;6vg1)fÃO~Yv9ծrgA[>iB.91B쿊 Ю^׍8ɕ Uj]Qm$Ј= GA'tB&" @ No88Bc 0*IPl|br6I_:Udݺ?#O~;鉪;m\{葴7#gnk\M;f^66]Su1X-[l[\_ǹYmٚMfMLju1(i!PoboK5 /(Z/o a1K޵6r+ٿ"&{6Y|$XX\ @ŀe4cx/߷Բ%31-HVO:^+mdL *[HZ?8dD4iYQlPO{=<]P;!b0XRSC7~}C\ )Pսo"f t ۞bO{ zNɓ_/Lc7ΟujڧɾW4k.67 ~isGqrog}b\1T_6l c+F;qN*0|I4]tKtc~"i(m| QEZ|j[EJˣtvӑAYE\5ӹui8س̯WNNEZ `TЏ-XlP/ki;z:oODLK Ŷ^h/YuU{jty{VM՟ˇ-*ys*)P ufU+ڊ9o]GWttE&&؉cg"T9=>q#\xB((/\E:doUTI%N@]Hң YNIvi>˯h{RdY,=^mF ^T?Sn}}3 ë/WcWY~jVKI¹}e7,VS)嘵zL]GӇVPۖ%'˖j3E7Xyf&2!C()@֚Z1PvSYM0Q[c$K^iK6 ]68,YPFpig۩mӀWo!%bVn5/g!˻ ;\f8'P/+hQgjZ@=L`LN,s <`c #?DWN 8a tgYtYpgggqV@>;(F( RIId$ 3"V̳ 3 #2S61BQV2[l?d-Yk4ûdw%FQpW,s8c1RuZ(Yv.FI%d >\y36MQ7/e)Ddx8"H+Ch[DA֠V 8\<ԋ¾GEģ 2iB@Hu^r!ybthqEt\^ pe%! ӽ3%x8\YT˶6D)CQ2gIRȣlqGeyx]DZSmOl| jf5elV ' ,SJ@|,"N+nv<3P5 :t֡&՘~S5yBBlfild6p<͈&+THlrs ,.VeYE/ԕuyEoU9[R]Lj(o<ϊ!\('=3:!Jю!I@#ӹeB[> sbE@\Ȣ؃E! Z4E]Яm)22YooDvIj}||ts4{bs7XLg`]KF }fRZ}Xf\n=i="h0 &-D*Q,J.XIB*k@< 鮶侢aE0bAD9 YD.;JH'fcNZdBM&Y|JFRZN30*` S\}K93e/; Z[ao~ qpry[o80_.VGB֋h<[A%}E*x?W$(uϖ'͛]L6WLS!׊6Ru)n[4ɎOQ BՅV ߮OoWӷ79?❧֎f YL[7nxj+ Z^zs>\Fkڛ?]6y&:hّ~􉆓ǰ.h~%|z1B5ݟf祊yDytۉϰ8u[/Qh?6&rZ;GMvj|}Uݿ/UKPccuq9r[ٌmf-cG?0/W$Я %~Ht4xkE61n*)ج*;ʫ_/>l}mgѶkM˹EH0ȜGPa(T|Zy@dNQ BGhCW8#vR%JE_zpx"^WD`}}|Ժ aKvJC:keB?Ε8,dKE:vY$$%7]"n+hk+r~׹讔SQhQqhHJJ \e";+ a|Tńs,EeW{UxfZT=hʟ}NIFW 3<@{GW99;y'3c]$檲`nlݙm=K27b`NJ?7mF,ߜq̩4_S1>~>t~%,qB$iۿ_sLjv ˭; glVڣ$V6IZkl|6ؔ ?ؘvhkA"i+Ҳ]EwrjO͂x`8Qcqviw;̣K[z!S:_ŘsOvLQ ۂ $9@Һ (]aJػ7F_)wCyQ(Mo4/킵Y#]]0 Ѹ`ͩ&J+}ƽ &)\[p|0 ݝj.\oM}g~{~z5~;x|:c^ W(n\h'{xU "b\d(>% cFĸ6E֊IEEӮ(zK]u>=GiOzwzəfN#q]IǼm~˷v ݂=xvx!Ig5mi13K +{''? GI?ď9~yR}IY.wzR2cvzO |gy3x\ODҸ*!h;]9JEW|I4';R~y<yQ1TM秲* RC?xZ=XκnBJVy^e.K%#sUPU!D d,;۹֝f.QnGj^|ٟq(ڑRFYM./e5ְNMVMBy]2d@erF*Z ^eҫ](&$&ZJJ9#;Vá$.~!S|Z{ FBq ?AVCwW$v_ ޞ2|L\^]m&yͤvnl箶z l8wE*~,5Ilޠ `+XH8wEqW$UPJɠsWo] Ƭ]}Up Oh&s\6_{S\8G@1n9EWTpKiq#j$ǂH}o4HJ:T©Fhaht鿾=)g5:m-pwS=aF'"gV2fMQfU))\He:*Z34I:ipx n8ϕgYr,;%(r Pqau뮗vVmv8 lg hHZYHJѝ޽I\v'QR5])Je++p]uTJ9f<ǴRos4]<ʧoΞ ]f?lk`;ZA$&4[Q0%eFKc&2!C() # *sU{f"dS{ f! ǘ0D\O-3ط&^C55yK+hI~ (Z^(M}MbŗkM)hQ\44'[g) [@=L`LN,s0B 68pD i 8ĨTp;mrς;;;UUr,L30HJE'm8$)a{ փ\Z3ϊh[raP$8DŽbs6h! hZvdfv;3ua簥&&.%2Xq>xbDdQ\L J|2@gmkn_X10R +ɜhKv h U,.[oe#bj jɠ8ia'B]E?*:`E$UIkGi6S&Re>u1ΓJ Ӊ3% *>ki'+NDӧ4̮_e׳~?YLA2vfa2bChC?ճxuyjORSNW2y` }L.hU)_)ΰ{d0^n:_mf_E1wO=^rgjyһ>pҫ_˦COUl7%m<# be[&i8>'lT뱻i!$zW2\\B%U`YVыgjzɑ_a3d0ӷ޹ncA2Hm[rK]n`S-,Sv@UY夘LF`0Bls*]X<49="5N˰gpk XC~8bG a9PMuQ,9eU`bVx#ƅY9h$#ZMN,q&VR).(OȲRd\&!k,6fhdm5qvC[͚q߹۳Ib/MO]c\ƷwjECtڜݛ ϣoƘ@.srR̍[`h;%MWn|')\- U b.p?= W Y%R7$FXK! Zxy$x |;^˄L{g\(92j%Z%!,yDm&+ǿ2,w)a.F%ީK`JexYME@s=lp3{qtQs~#Pu<]ͽ_ߣ#J Yn;m/]/V>;ݞtqWȭM2 륝~!1ouZ7Ժ}Һ= h60f!lewmx/y0(ﷇtp/O'!>}ǩ|e}O6wKm6o :nN\#opI7fpqGKDKx9e JD\UMU*oOL.r<л\bV{)j+j& #a}v}~+,xv+N( 0~t=m5`Sӎ x|T^$0|TE\>J*"")imRy}T^ʸ2\y_˻9#noޭTxﶽݯm[w A"Y(i#*Cd>&%'AN _XE̻$bOo./Nn?ؾ=[6~8[]ism-Zlk~WUbn KJ X8#"E#ɺ`%X;8u]ݖILRcS>F;jZ¦Zu?GMuz-\m)#s׍/to 5eBCkm y)JY%& }"ȋo[7bD@t8!7<29CT 1sD}fقr2`y#hsbuupɸ_;4lZ.֏Y\F`T\ގ&(B\ Q8 J[򈠙cYg3S1 RD,S?(o]>=ܣx+AVN΃oއhsG7UZ2z{_5ƪ#(ל>@tTN`M5AdD Iv :!:6xltr1;?kb:p?\.#zrU' F9T2It:gDYǸV9kdN2ͭ%lJy:y zkC:)J4:QdӚd4ȭ>wovr4u+خA>B{ҢU X?҅e˃vuIoVu;,כU{1{x&Dž}ofcPâ٧FGaUQbRh+S%̘(2\bvvQS"uWtWaqJfl,p]?tgo )sj|VpuA-vg 떨$eZࣅᾱ! 9w:A Y13ȴ+*B`%0$ B҃cxƓ9 Bi,ghcɖc&IEAH\#d$.2EEb<ݡ pUٳ-6Ab8Nr|xBcv4Eujs(}C E6\&2_Jx/ h^6O8,T2tv[o#+O(3i9$9>N"{ps[}ln2L,2qL ƝY\5;JM u{V3ZvS׍B5<gP'ʏ<8gK"j#d&+l4QRkὖ |TqO{dJb?9~bB<ĄJ ! 8 _\eG)Q1ΑP:;XY3MۛWZNRWvf{W3}s=x۪>Α'|' '$QBT?>L"R?5Ǯ9$0se\j{K֢oLUHݐHPHdeX{ybZ!c:U.G>s3U:yvCs~͎vغj-oyB㮞?~חwWw d\k͖OϠ)[`v-&@;x\s]xNZ.Ol/fo<@i$| vhh1\t5DA $Kll-DK6#R7gCפsIL O.P9ڐ\r1k6JSlwE/2ZfUq^߾ JIG%M"~XL (UFJ(UB2ӢGU6x㜰>I69E;Bũ6?sBpndj,7"KO2Ic^YT RmʃdO(20kL1F'Cȁhd` n*='|Lwˉb.FɼN&*jI(>եgݓ݂3 }(ه 3ބIԁi&D/2"K q5i@v1UY uPa6VVjVtVb ⸸yMyy#GCc)Ym]}}OKӜsbòmOm4t9;6}{DTʻ=jF(920I'aF-m ڠaHEZA+2ZhAHZFmMF,GLbd4j|u1}c ~ :BTK !@y"-&BF3Rjh ( AHQ̦&jDde.9̢-v5qv[0 ծ6;ںփִlKƭ U9FmňRw`CDf]f]]{6>8C&dȁXtɐĈA&a&H6~f┑Hu2p5qvÖ/3&Xm~jMeh8Xm31*e踳.^gO82btC)̒ [7U1mqR ˚# Hv&`ajndU 0RE"Tm@vq'_gYr]t`z</D 8LA#r(#R _! U.`v7Z/f`v$8'mmȝ 7D?ScG?.UQ'ymУ \`㒲B6AYduy;x!;bH I$^:OI!-GwF%|XC.0,pB$!9\C)ˮu`yhMC Eҧą`%% `]୷AJ4&!M8"@kj8{IsT~o%?X"6OǷ-.FݷXOuSyH㵬oi}rNI.'BK*(M(! I1R(R宒fZx8_dP~p,n -OdHʶr~3䈢"[U_WTWK/:|Ҋ7SV L-XyJaQiƝVc #"APF7'= pێ p/0j1 &'+d8p@AH`5D(8100%e0D+>Az3t'kQiІ" KRKFoKa0(u)A`b,XO݁e脦JSJ Xg9Wb%` Vy-V@mAFr~~}iў 䙕pf_nI>j.0F4!Iҗqjӥ_=iK-7S.,Q>d+l^0c=`16SDzcԻRL]LnzU/ t>n(85!'! bvx1YQBC%>30|.v~ 7D2HoLt~~6-Sa4HwJs&ϯ뚖QYum~2EL>޺R0!ӧYyyf_o~(x}=f99Q+?e)[oeQ6zVd{sipa-Ch]Km5C7#f D@Ii& 7 AOo_guNѵpNku]_֒DX0hң_1K;gQK,x,Sbno=%V觩(g~qKV%-i?]#䁮 eǗ0H/>}ow??w1Q|^wzUo2^~ {yKǞMC{b z.Jf״v5"Vh/Z_Ұ<]lJ[ŽX`-RrapeL ( O qXK^ ߏqOk󊋨g QS8j- \E9NK/vv\k6S4ݗGsy!9``Nt&Si=i*)9W׵94<0smCժ}jҔG[Œ3t+B0˿xM8Ba&uD9hK:s6;˜<-YǪMsy;TݵˁR:~u꣭' a2Ь*ːN.bJ!ϗaRMeݻA ½T_ptfKbkuIǎʾvKmX.Mwև'0 C4<΁sc̱|6=^r`gGH(򹖜N)bxЄ~r±~>Vj`GM<f*=TrĊ>~$Bf= d4M쫳Ww; ݧzH+opf5ݢz;zV,Xۺv~qk{B_jqa]]LUyIZ\ť 󏻺.?ߴ.{]张ֲ٪2 [c-æsۆ*Ǝ){Kj-侖Zk.dZ\PZ`ɺ{np [8[µoRˌ0Bn 7yD;6՚0kԖ vaknC+]${nM) :&xi[qz5;V%xm1?qҞ~+y"V " /0is[]C|'c:\o`VSJG4J8c=UǵL=ɲG=\t( +ܝ{ .AVc~J(5:E?'RJ07*ԞMї7rfYa {Gfl4 yW/-_W0+:%Or|Cif:u'SnC@19OL+%vkɦcm!{13DL٪,$p:msL'%gFShY,;iy-7+u@w΃;hBq@:%; Gf滯_meggBcJUo`,\ՙۀ6Py @u=k=lrsre<=CdJ1eJ-P5]z5Ct4p!flcR]@"H,;DW WU*d'ztEBKi==nMӆel+^=*ZͻtIa|Aoi"9@eu!^@]bH=ZR5u< J_ #HK0SAylۘ;mսVߧN~,O~j S%ǿL(?j/Wh%l'wGOWR!'zt3&F+^Ž\=`+Md5dAtC`n4]/]Qս^`so oYeP˕eUUK[R;W-κk`oPtY߽(o3ENuu |_B&fOaQR](B0UW_].!9~ hˌnִN(T)U>abJu7G ֪û}W]<ڢ"Mf>).4U_9a̪K~TV},HF@EE:X|X0wgIJf!f5pyΔWgtt#SBo&mL܇Og#7?v?pqnu0pzñ' lGvG +dj7!zU2X>>7 Ҥ|gNh>|YeFhKsL" KErl sMÞdI];XcV" BlJN4z<%1yaܮퟃZ'l Nrꅽi`"{&CD i9X{)X$y nbi1g;a -8Dhe&j*C%jehXᵌFMFS-IZS!W)CpR;o 2rSeO"sjA2f& 1 E܍ li/8 fPc$D`f1#wBlu$Fors((5 HaEc0f^a2~kXBt@QF(X`*`#I7@B)(FǨdt:Jä>  b2R B b RG@#\#v$+35.#%q8N ,+.L;(->1i9oӨrX`#gBc9`*q{6T+[CvJ= l:}NVmӘG"hv 62(cu L&P'0D:ZԼMHXS@y$ ,D8Qʃ&x<[ :/:U yiv CFPD" i4Jh"^;`C1R!aZ\@GSa 0x#&0L0_Z<tXř MLT<1(;z^TAE"vGJOf偭raO -.2pӆgZ{۸_!! (hp|IciIU޵IJld+c[6 $6g8s~uf3C  qah Yes]QbP k9 Y"*P*JvTZ2ɷ Fԓ*eX\~Lc; y [ZwUTiRS`Q1 ]Mh(pu08pqLcB`P&׀R߄LhQ0 DeΦɠRȠE $e5̀j@o]\Q?A2n@Ơ) uڀb(4VH(,@6{$؍ttg-JQC(]e֜G41@$5T^Ue^i C.kynse5w‹4?|#1diOԭw@7ۭId3 .-zLaw-~)v%[Y:6 k*%'ՓFhnVhL9'zx6lDŒؓUZwCP^: 5\Q py݈V]~nA5(QJj *P22LQ%fTiOdPH;@qoܪGdR6v"+TOۅ[QW6b]$r,JxyŴ ap`R'J-*FjQDFb1;z!i,WQYj0Q380)`cFj҂cZ b $_T4/Q40-l]΁?!]뜼Y%kP׮BT`joZajlJ؝5F-6R FxX@Vp*#-]A-l`ԋ@+Bb4S"qfd98Q=FMWPO, %׆!Inv57 !ⰱކ`ݨ>jE#Wn6˥!0r(:ff5$mXpIFi8ub鱋hIҚ^4 :ouhx蝩yd]pB; P G]5B}/n^.V]]$B*ܔM]c}{׿r}])e)*z&psz}wFRl&R./AN%d{KϮ鬶z~z0!t}7]5t^NON|Bgj+\^bEWZW{.fdQN'﬈mZ7ۻq}[Oޤ za:FکSe~TS# CwLJ (L*@@^J'r@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J҆z'wǣ~iJ ɯ0{J T_ZJ V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%_W m1){ s]<%P4H^ +c%TaJ V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%U|_@FR(\#E DG"+`%ST)f%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V(>nwX^LeAtכv\kVW۾IY^.B.1 ˜<ᒖZ.ZKRK.=So5x8t~67{o8+nݢ#~~foz>ސ툒-$[=dKGOeJ/JTv;Yk$v 9'ȭ%v<-ԛXsƔo6z Y?Yi:OMUzzӘF\o3E?- ~^>ʿVC|{Ot/yz^Z~Dz|yxjïC[m|q9]'Xwt}pڿ-9㵣m<\)tyzr"&j>&$S3R(ZpPҨ K{Nnf˖n1w!ymwMԥqc/EN߽?ig;ͧk86rQP7GĕQmyj^.;v_~vGl.f9|~^?u6d=_\jJ׷v~,V~F\^fo?6΢րTro d 0IJ,3U8V}WSAV3*6w{t}VH!lҵZݴ pQ3$[ģs_DfS03E>ビxsEp{uԡgǟKFLL*Lڞu8;YZ`Ol.?9swl; Cy}-Ѧ<\/;wI9雋/[[h-OQ ԴY4uR(V歷O7<~8{~arVC5d =glr6Fc9 O׍/<" sge#bb}oO~J~=RRlu9\TٺJTD/M–d: 'QwS솇f*2@OzI'V:w[7߈D#DbZ]}4>8}Y_#ۛM>5Ei9N"}QٷDCZ 4Rl}6\ljN'=X/vGƗҧOè@-5gmIn7{cV]IQ6c5!{&v"YcA66.^ l>kfpRwl1[]Ûgmo@YMmy۳q̷ HZ}66y7~;7,^-o?nw !Grmo`ȡlHjpM .[}FqL:>q|:Ko]Cl ޷z\"VfEXѶnFw\Њ25iNT)MH8Ft_ʘr h˕}8p.,B^G5?ՖZV; jMBWoPI:_/b)JhU؇^|^l\WݍF?qmSޏ_ͭ׼j6ܺ1Xs^u}{:MoFk%mf3;7$ٍ &d=,]A3ҒkhI j@(Lfg|g+7+ rvΗ/rK];pݗ=zG86ܯx0/>E׽Nךx)LOV孴b6ݴ59N<$rL=yr\EG&RKdq b\@VAO1ReM߿NA摤q3G}W.آ0QWkYmI- ¾#HUE~a~ XoBT1p 0tQfv|83 /)w'}| U[xQjنaC4HYކ*VzLDSRJ˼񡆵2O7n뉻] k\JGPmqݤ3nfb筬b|Hד\?h)pӳVWqu=Hwk Tac&Ɲi86@475Wkl^OyítW`== !Ն 9 F݈w>bI;D1G|vN]7ʊ!_=3zfX_bCV{ >־\ 80AZ 9%sya(YD~yA*\1vόY \qIC]" ;M& |Lܯt+nqe]٫L__9qؿNEy\ }L'ʄFP/Ջ$Ÿp*+B6iP$KO'3OL ח{Mkhݎ6E `W'r0.qqfиQ?'c]J_+@$hvN2ޠ6 JɎFA, JZh@)x iLfpI%Uȑ*4I;4,j SNɑ~LfdF\#&hu楶ˤ+ g%YPeu8;Y1g9S~#Q[Nē)(N$QyQF@BEy J $( ېT廘z" B c+/ךPPPik鮊D“q>^}nkwg%ۜ-M89mp9dv6~@ܦ/?\yGM  uR RQ <W^I4Ta$xёD4*gTFKG㩌Yp ֓ F꘭NQq%k3Uj//y!U1 E3HY=li N>Od^hhtq3N/c<_1HKH!\$@p [cSl5}AEUQ@Y00uK5[S6NY5H|G ɐP #ޜ%:jO|qwLjX/_b7z<]m*ƟQ*F.Ԑ2ZVEN5) i+|qY ;Xؖ |9 6w >f#10&%`i2JW2CA#$Mɢ)cqH\'A 4Q Bg~!Z`?/JFtֳJ&2HN 8wҎ\%4cdYV^uAHT'`#m ol@3^TL -iXGm(%ͫ[ h" GYo+IahHO.BJ?ޮuOq"]43E,;ΈrȝZZ)ٜl  @12!yDtApb[[4 Mu'ŗ,KK/ ԡKRgU(5sMBRB p{ۣ"!iYcy ĺ ArVkIbbJ%oYITQ'4 4ZgER'{'ǡ%vj uCh7\R6GQKM #b{BKH8=H@2L堬BXJQ@lL`2 C2ZX)wv- =9Gip^ĉA9=ř4|\xNtpq>J9]qi羟ғ4^{ ;?_OZK&_4웁G<ÀI>+O!c8epKvr\7s?|2.]HD;x-'zGЖAӺKي[4buKZ2KJ^8+Q%WoUsPY^[7O'?\O]Uc%%08gvoW$jmʷgnQu$YL(t5 tOU41Eˣh.I~1}^lvpiqFt lEvڵVcŤI%_/BzB+'?჋s~+تð8B|xqpJ,K ..[^B°8_̚q* '0bw%`RafKgaxyFD㻷]?||G~|s@ YGӣ@oo}ȩiSojyzo^o:ǖ!SemmsWxGTb$=hxQ耴hUtK.VѺ(Q(R{;k;̥ict:yR0\Ȥn;a]ܖ<=&i2*`!~T9ipb wԏb2êe{s1tgC?\(.[ʲi|+mh< t\])=Zˋ_9Cc՝ֲ㮯ywݩ }v:c=vh=vtL[L(Tf^=bi-ԈriY# :HG)]u=F3҇jɺwW;}qVӲVwLA{hH zf~p>KnT$G> FEY^7$bj3'ߦMn }7H_O㋛qh巨 FWזέ%~fEZw/mi=%]@?i'NW$0h|Nqu$N{'0 qR*T^]jxmԒIdHР`Zb't_x/t[ km}7gfu|" ӇD^N M7}IW E`Uƈwh{Y?~ |ZE"+UcX:E8Rw9]!M2\:Ƚr,BJ^R\֥Ȳ\{ڌ\eėe2xhP98s~й o&63LhϞ]ɫGOq^Ө>梺ꓦ!YT+-Gam1tXIy >u\Skci',D4 Abv$իfTcJd=*agShX(]b8gL"YY -K$J)95١Q4cU+I <~]GdU>kCߊʎaE(>(1Y2 <5YۘKRMVdjpINQ[!'uZESn5mQ/{:{ER5Bi풅 JQDѝ[o`*mYU:s,;A@SKpJw^ &imP!@&%M`ΊtV R57 *=>bJkn3(Ao:, UdmL E)0z5II& RΨwOd%YͲb"*VXWl.v|sF3sMu˭_6^H XD@EWmf`N=fFhh ڤ!Zw%VuQLm6,-bXo9Խ.P >ŧ) z>Esj>Ńz|ay$]>pI4U`,$r xOHRs&*|N٪O$LZ{-v4RT!Ou42QǸָUDK VH }i'O<4ja=ǻm][hSXzNf|Y[mXfU}6$7_ƼfBkj:d^g;y,y*}5>[4F+.3*QWR2J{uH)uΨVkFeޠRJ+ҥlQQWWSugW[TW򕺺 Gy9}Pqp^o9i\vT=ZXx3c Ãvq=B@rusPlj<WWݢ{syRbo,鷟~Qo^- /pg`?Fb[#>r'AD ι+YR161b*ؚCi(z| -@ Lq>&zX]7Ow.t`L?ژtHrQ[ZD OL3l<0>j~&p9)2|HԁG^hl!F" GG03&vxWL[7%hb-+eNeuE]e(իQWг3sZWWl^*rJ^V]=-QWCO]G+֫C:}u]QW.VSWWz'pQQWd]QW᧮z(tH]eJtF]e3 jz*WWoP]U{HwF]eJvE]!Z UFtޠDr%c0ռ+ AN]]e +E !C*1.}֫cG6=dz57<4l1Q0j!P&YP2Vi4f _Q3"ZQrg?EI{|*GX|l6Fd3YH +Q"RFga +\pl$$_ӋKs-MM&RXNF1jh S^РpNqhɂ 8cHN)Zv!{Ё *PD&/sBGW;&0R8d!jg^0!p<,/vuY瞘ˣ8% OC?)w=8&pgM6)N#jygR ;r7] '7_<^s8c5qr]9F}JpF599aH9;Ws-xFg Ybl$Z@J8j=!HtBytWW*;d?Uh矲LNeebC YoWR$^žnۼJ}*;+AaFoV(SӌeӦk\+.p O+*UI;m=_iNMɏՃgޯ `+0pFoY횃EmuVvQllX5]nmFɭլ /8CO*={1vpqZ]뢑jVXIR3* ?OU|W<LK ]RTrqa'(sT@+]Rg'9IAX|e}4^a9"6iWDhDInp\?N?_>}xÏ߿|??2ˏߣ,z\ζAeoGyUWzY5ַkhPloP6?߫R!9aD`J)'%¯\#|i#e$;G)h꙳DK Y4AR1I@.vs;ʥfϿO 9x[>r_ICETQ-7sjl MSQd_:-qv}uV'(6c} _iiG}oa^P|q 1;'=u5{FhUBuG U8c _K<ѝŠ,w?(vպn|!,p++"@*պ pQIQpʶԜFӐ>o:6 ZhW:4A{쑒;&L"UyRI^"3|/7"XgN ~dcxr3oO,vXy6架;,zvQ3_}NmI!iP14&(4T ãT #EҌ<'= igRʃhښ"uI{ST)ƄJؖgpkp<d2Kn.uv~=\dh[-6EeT+z,KoY^†ڠEPҌhTRFL*Xȱ e zmA)aiߠdBq@%+CllŜ(^(Y$:* 嘴\{S]hch).x뜡DB0*QlBȭh;&[ihPT'M i>< Lw1!%JgQ\ *%`$@ShUW%(rD=*q$#4e'*N)}W%ۂi?4N 7iv\\ݞJ0wofx6mcDUn[[?SRvMʅho.ll{'@@+I iPDQppJi*pgV5*#ZX\tHٻFnW|iRtl-&)z"-_ yH۱,9l^,{=,S;@6(y<9\hU 7ڜޡ\'sZsFz\Vb0`^pբnŒov;@{O؝< Gl)!7@Q$!DZABy"+&B )s2uC{!) lJmpHIf]8c,[ϭVsF0<Ԯ6;vEm] {ɸRsd9h5p@sȬ (Hefì1XŁ21CL#c*ߙISFg՚4 :T`BG֚"f@7x`VƨbT:ϢκTzRv~= Hѡ2K&lrƴI.kht#L-d`7d"t5iD=M4ԁpqtuH{fɮ*pqō;ZIIKE\6LA#r(#RNJ^! U)Ѐ.~gq(xv=@u>]#Jo{ uXx=>7x?sr8Zknsu9"* AdXcd剙$(ЕKm^hkM/6Vd9&1ҵ/|r6I\H68ƜDqnj6$*V;wNl4xHL4j.c4. P$t|ڥ爻,* (Iw rjHѫ@C,6]/ς;;U[2sp"DfdOd=Dä HMU3ϲFHѦt )q ݰ]M8&u"+ f!I2de;֜-R$ڗe0QCI*. ZJ Xq>xb(Ki`Xr.FJށG &Pn8Qu4,k&Qs8r9pb`'4W$#jK[HQO>PjT+ǃNt҃zg)5Q %'Q( "`6&1!ZzbtpTL\mN|!):ff( C%ѥ9x`8ZY4-%vW!qёGzˬ#Bm+9䦡>;JզȮƧ*o B!̋O"|IFΝV2b!xUy3;fȮ֡+[7Agea~CEH9xOqi'JK$!ap6l"lZJW'v@x3sNw9 rh ^=PʆZ{%@ͯc ,"V` uT`bVx#ƅY9h$#ZMN,q&< n!ln̚1ݐoHLzss1P7%<9dVq$`J #ku(< C!ȡI(!˔w%ʕ hZ%!,yDm&+ a[iމekϿW: 'Vw66UZ숚Ab.?ntt(zJWe؛mTxW{?ӕwنyPRy׷tf;o3 tw<ʟDqgs\'WKee^Sŭ\An۠H[%[J $(OFNmk@hhB:1D~DN9^S!V1&1<,;<H'Ns/m冖(9r3WQ'{'QEnә!z韭%j:n.~Zh wE"+Jh9䭦U!pڴ/"k6: $mEv*3TF4k\b.J@+Yؼ8x9_F._MWS^q9 AΣQq59[^ 'RD(sВ 5`cd2k,i' 5ƌ6ĉ)PY * gUL'À#}?Hқ1xEIg`x: OJ 2L4l1dY21 Jk0! (ISxWSzⴐv}DuL2d,ɲ(3Bk8@P\dz2O15@p1r{Bcv4Eujs,AFxÏ~Y[6n0dޜL&yzufOGfP,~m!Ԙ]b im_51BSf{~~XDKDƔ8k ;9^8V'5*w|y~rE ?crYqBꝍ/󹏧 }։1~I$0P'~UMz-`㖯MϫJcA {N <ԼT\0Js/!C_,~"|tq~KRWUBn?i F7-Wr,/At)!o']蜄@&"u$˰*ĴC} |U P8Li׻ bNء)-ٍ.ݍsWc!HО|b %f'P̓V:a } Ȩ1-֘Ǐ#;>MuMv0o#a,wwg7ny: F?)2x<NMh:IQ<̟LiWꆘwcEXxliKVە ͮˆa8k@.!חM [Ic5wQ&4\k- 5բ~H][ppѷNNćBЭK" ̓fQDۏ[S52NvwvGkjz-LgAbFGg/7~KD2ٻ6$W=ulyg?̸{ ,f}nc$["$-=C"TR*hۂ#/2#x֐4^NT8 ~UK?rTKK!Q Z棃oyRӎ=^El+ϳgvj.gz4!_ n9˥4j]XEx <0?.LֆNVA=TGVJՖv?-{~P\>eK'q#TvR$w' %{>#DWsiL2&rxXr*jzp2;H{,_S uKߢoQp:Nsiz2`|qetKzP {-ɀֽVVL'UkܫcߗD??=,óƵgϥL;\4ARFQ>vW :;=a{4}墛sng]/ꩫ/, &v:*7oo|ΆM7 66?/Ʌ9IIYF? mQ|NgoZc<׳_>O"y82>->1s~s(NbT#>ʋ(e*8"\GKk)JL$_lPNfgoЃ;.\j ;RZ+AQN H h[yC:]ױ \ c(KuRkۋ' "!qDÙD QЖ.g,L~1Q^>hh-Dte3LNy\bi 0sR`To '4نv3%SN4п\%y+_OYm'hy85u%4nzŦU˧NqP\9ka`Z@"CsQGϩXt` b۹J4׳!z9jP|Rp%?HQ,4@D<RT,Ikƙ .FjPٙ҄%5gIAL21j9 YԳZ蒚>i;"0 RPZSH ]ph. (A {"#^J"KCNc`OkAQiS5w  B+ kD""=*8i&gA.=0x^:A0ZN p*obxpPፍ.XBp1",&R>~>)dφpԁ:Ol Id(Nݶ)y/\EJHN+O,#}8UfAƅuEȡ' +ߪv RE] STr$g=S JRCĝz(k0VsWvoq,;Vo1_6T+2N^62@"zmcfO=M|I͆d-*;i'B bEcrH+GdjazjțK1#4cGb#>/1o. Cᓴ( 5M%:'*ALP)p*&.5h_r))z6̙ #$c⩣`9JJ)֝-tBw 3jg- gU ~ir;=C-;%KCWfUcOFz^^#y2̀!u%ÁTG^ ^^ ^NH<&kACr@AIV1-=1t0$S*sZT21C3b NP"RHSN@69 ܂t*'źӲvƒ_+d|Մ՟zx{AEjv>xޢb;7&63^M/pǓyf'>YOQ'vm6b#[MlB~0s)cWU5ەo/}fM-߿奼-o]VOwhg+%QjwpSv /KjϣQo9XjW۞4l[t||.not@|Qrf3rcWckZ-u'AVçAzƵҌTBV۬r3U>H?RyL :Wv'H?/ttZ%$z1D (j4s.YDW.A:$ Pc8)SA% ¢qeU AT8H/֝vC$ mQb`5Xq@M =#w>0'J+W«k l>K>jrmGBgr wY ˒URiQ9I*pJJCt, k d5@(@Ni(5X*6 JRD0( |ʸ}U6h5_;%*}ƭɇ!fɤ#ggNKMZ۔8GQ;%0xE4:eic "*-S"y+QMQ%Ƙ8qC:&/aPD3%g:vK֗H,֝JF]sj꺤2(dbب/TG G)[M1~!Cԇ1K[b|c]\ᓿn ĭ1IhU M%,'[[ۆ7l#9(mE t|ᑃ`DUC B!f' LJH$`BiPJuep[L8Xs8?Jy޿|n>!_.(s1*yNT9GARGbj~98}R Q~k&Y5*T>^ڜR9,+B8L\zZ)CTA 4rA:bă@hB#PkLMMX(O*>/qP[y hpwG9]kʥqkۇ}ٴua45p8TJY= ﹰ=q֝E-ĐMzVoĀ|^-~ק:6t3?Lrj_n |Gft|3Dи*xx5JMΆ8ۻ"C+`gtlbΊ:(]txAF)i4pLxl|)C^vVu?DinK4p᫵Y]0X[}/d{)2P7+p2$dL+DqSd:;Zif||z\'&fjj2Ϣ;5KVz$KZZZAU c,qU$}Ї1s+79%ǣ"炈0ҶxgSbVFF ld(CW@P,eoq))\Z@9S HJQ0Ĝi Owug7Jylt,w^n㐶|yHlHW@O;1RBR7_X#G_(0Hͬ]D*キ$3<d_[edc){P_#D<#DK>\f; 4>8!cm]ޗ ƥH=gZ >?SmMu^kSf>8\2ϵI,>˥[ƨdJ*t yC:I kH%aLc?Ŋj*kmHn2R_ k{ A`N V`^DJap믪GNsM]0Y:}>c*ZYo+yW`W87Lܜ{.;ehA>ETWtxelWyA*'Kk&(FFzJdȫLZL2)A%9MRPn%KPF&;S>"QpIU9[e4Wdxv4)9:2u18-H\d5=]&(U@-СN {%NX rҾb&LA O*Z*2̈́HD"<ĥNM0ciמklC"WUg˧<$d$FH%k7D{Bˆښ8 5\Ce]嗽YҖ+՗zP2]&$+٠kwW֌fݰN2y**ɂ䑁I: u^+ DA#2= ёhAH"M,+v5b̒ 021[/38S;t+pdmdlMWi [ӌmPawؙ2.̸adN|їQΚgW<&O2GR;"d베X82OA!rsC8{. &Rؔ ^C4FrdhRf`ynmp\}Aִc[-PcnN2Bj.L9 m] d lfmx6.xTC&fȁ !Mđ083d\s­k~J}Ac[D4-#C41(`EPY!#.lAfV"gL 8)fMc4@~#ጏ> `3dh[g="~>M=h>:[ӒmqѶbF<_m*3$@2XVE " -u8Žwv;<,yt=FiyIp]3e? Z~F1tu"ZHzX٤&+t@˓g3ݥ+)[ʷlOR9?w&өtmg=URyBmrP& ɹJ'!2,k/yYL  Eɥą`e3X8mСR%PXrJyM{[g,9MF"~B)"YhH BjZJ3=5725MoB$f e-7*ImZjgsQHR ,B !/\E ΃"L 'mFBC15t *`FJAaL)L6A&fEù!H !i~kX+)I`Wvu9 1@/ieVXZ&`h4&ku7v,?2?дQsH, e7 G08 2@-dTɋRM`у$T) tJ)e+ *ʋ ^ LnV1$$QPj"vn?~N~~kr%ZH~{׃*4 ?S* Ji!4E=XzAˬG7m]{8 vljYѤQ?WzD{%|yo8 ίB4,{?]fߝBe2yc%}%+( 'Iڇ(0*t%@{U:yijt/]?o/y <~C7z/h:laR!Q"(@?B;M͛^=47kZ"lдnA|vy]tu5 q@kYs,I?IF9 ; -VqLƘFHwq\Fw~SfR>|Q !*vm' ےg0,DWݼNK:{U~5*iyILxkp~ny|R1C\Ky`dnr@ R'B.4N%\YX0U aݎ/0|  "0[Qgxǝь90*#46qD<0-qdz4T Q32usC,ymoӖ8/uP9G\.|ݖzUBLUB;z&vk;cG̝ Ii#8IhEYΒ(URJ_8R̕`\cUy+.)WTme-ӚN*A@j|uo#eQξ" $5 XZNu ]_.jA:ZѪj}_GDa8VմZbE 5l5%QjK!- z)^[Ƕ =79D+1[ Diؐ>E2alCW# W=נ/[`Q7h>B2e2(wLaҨC:%2Ƹx[1){'WM .^f j]rpVοy&XvЈ,b4짙g3v\M hT<@luI J*qf+PD&a|U*֋_o#y4e1A52ep'mȦs&NeDk%E#srbYjtW㾂5cov3Rca b}ʝv"`P=ͅ*$ +6HSnrq`ˁm)NSqy07"n^T?QWJj@0ȨD)4l/^2c&8(̈́[ ,ypP=(~K[uGl>^gF/J}ƻOd%=7<ԈI271W`eP&a}D r)E)rñC By0tt3Зo2{ V Dj9bJS *͙2g:w7 eCJ?ɇJB];o3zmVf d֚`H&F%cV H[ ((Eb+'O;3zCVG"a :$0H,b!uR x# zyi;o}mԢf7mB RԶgw H-6@h+ a|hqhu 4quuO[XD&[hP0NJsa<79O{6n%|*%7$[Q݊G)"8m=W6 )LA60h`F!+T}]xjraIGfUV~w ?�BgDyf|94MG rNó*o [Yr8Z؞\PNd=Ce|6bSջɅwM*ݚ-6[BDt>7*EasH6I @ =\6nJl_=喩R)?~O=a_C2YCkzF~U+EoHQ!Jf[Mխp6Ti " ?y`|?@ƒIJ+rNk(!:/~mbYS,n<uס*Qdjb'O"%6ʗF]@BI-q!@0op4(/fI2ҳu-Oyˣz$g:hCR__&ofCtp;]GJ]:t#8jRLudZʹư,Rؠ,;O#UW0ZeN^=l%ŔcJ-D(R6 ?ς_)v S57-~ڼon`[c9uLlJ*Z XHa|oK&2]ģPmK5q۠ϛ$Ja#&d#1?j;4׽z8Aw~9yTӕTV`<]a]?Ywwլ[Ҥk`NwOūZPDzP mn^'HgQBQKeF)}vx_`>%MŌ?ºbu\ 嚰Q\ݽGptNT4Wԥ<~%?6sBw[vV4 SLu u]vD]@twl\vǛiMnv $/%XL]- Զڏ0v ocqU&Hܑ1P=nh_-' r8On'hrqw UŽSNY((hǔ 0Γ!`z*dJN  :kdJ1>9VG*zA"52D ghM6ӳ|%08"&%'x`S&A+l$Rk\X XxYVa>e&gÍ[dt־3zxz镩OIβʩeȚiz dc(&k8%L9(TTTTTTTR=H1h*ܙPp:OVff%U 'Ü%Q [ZcI"!I v!:( PJWXP51&] SBan V3lA!aN9&r6y x:2,"#aREbјK ^\蘶 -mΉG%_gDwք6 TIM6 @mҀo бʠ1EL_; zSrvaI &%G5`ɯ!:MR  F睶B虄gTk)6BL!ؔ+f8;>^JKScL3l[%/b~&OY~Uu;¿B2x cd3& %yk"Dz l?ARӘG"h0leF͝QFGR*:d2Sp6pSF=/@dzds}QN@ԿlSFdY~&R3X~˲o>JƌPLU =8kmHLAUUcf SKK*oZFWHYtaj3jX$"﵌87tEj4 ʬfN}DiKGpz5GF^X= ϴc8&o ( jTpq]`$9J1,,VJ\fB~B zNې}Yr\Lo&>|즎Jco'>qk]=\TGWjo9/e4+WQ9=b2@^P,Q .ԁB ͥPDM@@|V(ާ+lPG S\-۫)CP1g*xa68̼xgXj/f8Vvu!w[T5ټMҌFa5slX2 G(5?,Gbp3P1(( 6ؼqnFː= JۤB:`^D) Ӂi]I;-89y,\;vʵEf-N\L "T`€ctNakL3 w f#E#R^凑 iU3A3 hG^ 2(>M~~X]6#fÏC9#8.gHQqa#V:$Q/PZh@R,:/8)"*fZad`(Q%) 3Ij$"et_gA&:aM!PvXaG4z+^r aDX$)/!Xz*G":)B99:3_p b @{юأPLRcZZD{PXϢv1@$9_+JwE]/U 0NcƽN]%P*}j>?ۀ/mxu1;MmzX12`)")]JA)g*\M.gݧ(85!!p'Kp ,]nE¢07ru~|aqx[{M15j`etjN4J a)mEݫ ijZrDrjȊU2XR50y^ܚSۺh'|T4O~<]̯֕ hNkn8UTvA6nʅ/]rqGa!m$摮 C#a֑eHu$ǓGM"IOƣn՘|:*AG&mmsU&U:Is@&3^Tb<}Ťqotmk5T.ӱ>\ev^j_prAxIS;4$w?)׵ bw!;*Sk*զDžNH~ŋgx|zp Lųo])3x/Rm P`~yL<zMzhM-KW>81.FEMq^ FkFw7}ODgo2v[GGs}^ A҅k{O;ZMP7'yFEY -.`,'Xo*ȁݩ`FyZO9y*{tmA]nmQ(1%Dg$%pŒrRhP|VG7Dg'CXLP{s¬``NLf>IDU} /ɾWk:ſ}v`}++-/meSRgZHBcwW?Kq@vsE%`QH$+艹!)J"EZI\ص-q=ǯ sԱlM *V:Yݒ[w5lqI#ET(CXԤq*Kdű?8W8dT ،P`Cm -1g"a+'oYq>D z xpk: QA6u$4=(Wqs6;"3&cX9"X,aiT1Yʕa @#.vkO|m^mym-VN^tςERڨ1=dK}16!{jF{cCyǗ۞rn[;R==\4lUBc AHh- IrV.qЯW @2 GYHEƂ?ŀIc6 ұMNK{5A1Q۬24ܰ/mٕDl=EgϪ4N祷JkhržC E4JhB)"UvAFn-`*tt?b-&UX3r8JBZ|o]~Mk/Z^=4kcHV,onXtyZlTu@\Uy!6`pۼȭwno<<kN`N\Mv+Pmhبܸnny=~v699ݵ%Vo֒*znZt77۽|"*B9u m4Bp~RI,P{O bc'4hVQT9TD騔%"M(fX//{0J@D&5! oIE"'JθN.׋41/˥ǏTQ ky\)]-ww_ٌ( ݧ8KD뗓2KL YFI~ZUfwU_ K*I8Sс/XaˏQbi!9ƛlsY g-'=sY,NCZ8OIHZ%'%I @gJBAM7 -9K (|v1B1^ _|Ύ?d%U|pew˟}aXU^ :@(bDVEFQe26J@ɱ Wogz 9Z(Vd2%HF`"h+F%Złٓ9*?P7g;`']RRQDQIF2a`X ER/uC"!băEpU h⍤]6jZ$% Q;+J -!e)ڶu(B"Z+ʨ=2{j0Hz;-=n1wyKi;M ,'>>j..? o%jͧtj?41'ūL`pQ7Yi^y^g۩/'askdb0nI_K}KOzBT􌃚/n`rzrqfK[/L_vb}>j^>bٓ[j_Q.P͟:bl/I'{?4^vU|FÓwY0,z1]yc;?ÇZ(<.; -d35oUV+UvN$j,NC,632,R膍 4$w>6 f˦d &K]6M75^ _ @W!T:p\XQMX$Iv._w%]2dXWB[SbȑF8 AMZPjSh2BMXe)*D1&Ԅ>XEV&2H!8: m8ݱ,;d֪ݏ$՚4UտEOnd}{.8E_-J;Ĝ?iRxmPbQI4ޣb Cr炗/_`Ş ^꽀9)]T.D谄d>dJ(gK^JSJ1ۮޚ}sNAJ3B@uPDJJ4!BEh=ކeFN6`9o.F'^ej3_</+&kh2ߵս}E o՚xڛ47u>+ ]ϴ].mk~A>:IF4T]^k4bzuíMy9 .l~t5e-ncƛ;R,Oy0o9oCa.֙}uCdzQ9XiΧ?omn繮oۼ}ί>|&llױQ(w|t5\rM⌆qVXw=w(du F1$P/٘7Ԅs%zmw%z{a 0"e㐤AQR$ΪX~(1fl4~@CP%QӦ\X|PiEZ",}6 _3b{ozxf,xӋӓZm8xfu3B|53WX”l_K6Xë5Ukyww3#?hAY L*gđ5koE?tX gӷ^_YĶkI{p4X&>ylbEoA.ڨkUV!EԢIL™X+Fmí] z<@gc|wOX1fC t2a9wagv[m={ʥ{+[b|phq.w3b|r5i\{WȔ~lV%vZ~^Λ4eL.'fcP udѠ&KgК\R*[Ra ((M >ڌRZ3H,C|$0*b*Q>l8 qH+*-f X0jZoWΏ8|=`@fYx ZӲnCm33c;Wu|؃U9s^u6ֲ'u܇LCxx K..oYLz!o]|ɍ鬙03,Lg2\4R:EZhc'fq Lk[+@]bou? s{kϨe=f3+&NG88*XgmZܣ5bp4e '@xܡϯ\畗m7l|O5FedŧߟnڳAJaqJdd m@m扲2f=I%)͏o̠?nfiafwo*| w?Ydp5Y($#c YlB} ǭHFAr^Yl0QZe2͒BJ$ NZ4*%( eLq[IAu$uq"YO}1[Pe D-xņ)N N( SBb #!8}7(JxW"GY5Qp&DaV]1(P?Qh4O/(SyQ<mͱϛTXJb c}͊uX8l-n/7AV҉(R.rNdV%dEZre 2zgv¦Ug7 ):-j9:4}VYJq0I/DpP=q_?k~؀mLڡLwuvV=V,TTD.JEho01[YV?qeӺdJ)mp< `lB:kf8C A1c"e&` irƫ@ )= ^m..90EqEx&P t2^)[lDma xLs<X/`t[5 \ ԁsHS-yRxgdfN 噴Mb,M47s|&@s@joU&Iq*Xtđ6Eb\hZPWg5=cɩbh,.!`+*dL[9kP s12D`"!KN\Qs( J[2 BDJSEl`+ XNUgdA d*W@P4WL%Dw@6$I .O1Tܷ ` ƷۻfҏCLm]LՐEh6F82=\ZЋ:" t*E ]J@_&C,C޴Jr VhGc3"Ъzl *N6Gݩ`6uSΨ D#+ Nq:H K;,@4 g@i4P?z2Z;hwCF W7 !\S\pYT2QN$Jlƚv0]niӖEs4YF%4@@ěW i(U95MeV{CEXY' -I@@ ~^3&T%TF`&t.YSiC:8syy:Z"贚z8M/eItS@0um P7]p&Si3q0Mÿ XeOZ4kHZ\ Gɪs6A(Z|ihMiA@SB^o*(9Q<@nxВMy!l)QR(]&T*PݲFrcP*4޸hMg*R3GQU9$ZCi2 vrda 5@+y!6ΟFpJ7a"Ilc`vN9rIbUp":r٥Pik,lJI0 ]XuQ@[9%y]H:V]K S D8ØYE ^Zܸ\_و8Lf3/H^{lt+/SUޔ<#ԚG>#3S1A77GoVki=N+nMC72`MP{oo(&ߢ5B/|Q䀼%_uf/n`-aE:UrEElPy*]b6%4;dG^|/%L.)B% \E(&߼0ڑ҄\kWlPeZҙ~Q듎& {Wv#12ܞ޳=T4:WdgIF`cKZe0! e5.M iGmaGm$5;.gsuPw~˓ufp8=O'P%_-|紵v}m))*ci όhFsh,&@r%:D qQDn-f7Df܎J3,lB U/y~AxPwMgs|Rohg59EMY$dզbsxCDlU}l%E]Ŏ7hF;ȾslteJ[Qj-NƳ@77㝭c_V]F: !b~"|t}лzA>w}лzA>w}лzA>w}лzA>w}лzA>w}лzA>߮G!;h;Z>Np7n-iS \sA34\/T PyfNj_=b{m~982=1o,N_NMϯgċq^1+ vO#/|3єZ dncHCTiHG4XC</Qxxhׅ'z>@E׳ȣۊi~=c4g\2qx_#0V$;\O˱m +AIq#C&f]i 7CӋheQYS2䒇0,ʛq@uy7H6}z#=ztt~rztdC>:?^;} ^fY ԊW/{J80?4慧W5x-.vm+e^ A-~8s ]cԝ4d6b6Lj{{x`yI(WS!H!P3J[~[Kpk[oW?gB]6u:#8p)wS{eݨg^/I2GΕIQAEۙ4AXڏ@cPeYuҟ={oxKURiSʊKr IrOFҮ,e>h^gU/N؋V9q/;?}P?칛nRy/MЪnUL|Yo7&ķvnM|o7&ķvnM|o7&ķvnM|o7&ķvnM|o7&ķvnM|o7&ķvoȆL|01p{(&{_Ƭ~&B:'ǃGSml}yvFQ'=Yδ(Jl}%G'困UOi:yAY ,2\ Yr! qQ)Ħۿ[~7,o? %0IIBICV{RB(LA3ʟkI=a y2(իaV^ O a7`k-vf-j5~:?V˳8m_rXr:pyQFS~\ໃus(a;xfM0PKډ5rT0R0Ƥ) ke fi^nQF5 ɉ> pw::"še wQ0FKB ϪbU/ ǜcDJd}<27tU?}w wq<2 E]Lgy >!9gBvM\D,g&`*'sol_>g:¢P^(W)G,. H1ɯJ:ɯ^,EKWE6P&d&/&^1撦Qbe[jDp"S֢B1B+GNxpNIʋ} c &PM2d]<9bp11X}.ށ)wVmWϿpS2*.DHI%IS xXc3$L%ΰҨiU;!Ɖm6_5[SuG Pu_:UUWu_:UUWu_:UUWu_:UUWu_:UUWu_:U}:LE,jjtX^]9y?h%UK0:Idz5{#!ƩSy 5 䢶eMNdz=HQnG֓xvջ)E `z@3U 2a9>qN%V2\y#-R"פ>HlX:!R3I͙` t{OnP{X\H&?c I}<"+xk0$~4J6'^1HP' 0l5#~'y^jAu];λi˶l/[oK'p_?'Ml#M67ݾ. I(gS]y^o:v)B3%i8Cc3r;X3C g򧡓/]=,OOC/p2s?m`vkMz){B }, }޾ -Ozbʕ#pks»\hO'ʹ<# a"'^ a P"p@F .3+zk2b?u_/*p#VUXVVpԈ+Q-SiTǁTpWsm0Bd7Q;䬣'ǎ ^??sI)icQbysX ٔ2֧XÀ\Nal.U"WlWZBo@٫68ȼoVʂ`F + HHb^' }}&>qCZ h\3YyinʉHPI˔v:s8pq>?Lػ{o=Z+"J Aczft1}kz ܉F=ÂqҜRkr֖d9LbdAk%kZlkه @JHMЮG0nG+5/|Yfi|E[ϗW]뚩|}EyRWt9jZ,$uVS*3i B! ̜er™~.\4}{\#T2gRj)svgf=WWlbL? ʼnⅳqDFZ5^&k]{o(8Fq|wLњP. cb)S ˓ں$Ij]Z_{>ގwqЅgZO]rcRNmR$˼Z'ȺYpEv7"(l|bj|;GHB^kO^WQ F͑42Oh,O̹A\5r1cn?(6BY:䗸z7sỷؕnj 0{!vw<͝K[P SOf 5H&<,lF1^־#P $KFEj""19ēȑ AրվDb͞g,GO+jK\O蓰ʒ(05B)Ip R22T115H1xhiypA#ϝ;trՃ󲄨\F=Y`9w|Ty$4&8ļ/$7./Y4"!5X0ထֆhȮWxWxy$jڱR^xb1YkI.01HJgR cԞcR*RP̂XsM|[߲9;}F@h׺|} G'&4B(7vdSgivKYxeO8[ܵ6YE6UZ41^;J#(.&7Wmt^f{l*JhݽM+JhPBa8lovub=2擏n8x3ɩ/g] 9Y]OW erCA;}1rb-/U4;c-|*eҙIE1"P)b"#5H ]ͩ{ {||`?'cC֬C$ W$H2CKqϘJ>;^4H/q.N XsS|B[˛jȱp@L03w%` S Tzq }! %wFg"Q(c G#xd @*G>,Tూ@Xsvm:~?%@ 9SŒfe辿2YkBpM.smrEf&z'9RQ(KFb1 Aȸ!*,q<]@#RZX" q rs 輙W|[z G(FGτ8X8DL!M(D fV,z4j矎7pHj[t l]&?2tlifn[v; C;TVPޗ8؞iY sNB F՜9 %S֏l؋×}Э:>(F'JzEԞy`Tjpq0V*!D JHL-cRS^"ŬNI@*:^m^l;+~^Z3|597:dD""iN ;"{gQ]=‹x@x  "N9Ÿ;0zy_l1qA Ȟ[}: KC92j)Dކ` >&lOjŬqa}.C mkgґ]moLtFK%x6igiNj߸ ӁH-Qk}P+q< i0\X*ZOM|7YŞrVng <;~p<o0aЏ;k{ww1?o/aDQRi9{e8Ŝ~qcuuƋwW穥:_1 q F?EۻaofpMZ qiG{!E>#EzuuֿG::o:Ka6|k[V_9ZC;`p'h)Nxaa{Ӂ ݁u^M3ԖL~&LC-gh=t6ng{8WiM0i; {0S$CR}_u7)R-(qEuH GRStdY)e:&Pfݘ1)yM&MAt402:?Yvqg{LvPd?y u'HEAF8L'qE ^$3Tr*1FP0"Pq"`~bO/ xx0%8"EkAHGL:WHD #@j>m\;Gݎ7nWO(ψ O^Vr w"ӫ>"%cG(qU]ڛeGS5!xt@K|JSjx j4*!|T?Tl}<\>Egf-α Z",7.'3J1l:w7祮^ VQ%m)o)1Yd`-Q%hJ4,ۭ5zZO*YpO8 n U o+?87HS^|)m_ϊjigj^cܶlxaGeh}S51ץHG(]\_fhtFêPAGJ(C"sQ;RGH*MO<49||O-S nb.r# "1!8r )YHt^",g"Dʥ.KJqTQ( "RrbrX+B"se+5 }JNx-qA29}R.X_;~Ad]жK!rK;@L%1kblG )Nm6!N=*'zx,9=xN^SihcJr N" !P8yd2(:o;cƌL"^ˈShnDdن ba:&R5u|k-!pמwpGv}^N7;L Q;np'Eէ[tgt FWNMBcSq1a1uYΘ[P}jވ߹n[=_5ΑcZ0L|;M.nƞM;:n|7ʓ/ڬ;N/nzzĺpow痲>_s\zalsM(W6ҩU{L~M#=KPip)$A%㞔КR" G 6'n듋5çc5Ft쵒i0x% O1\2G)'a#NX7hreq BEd7zf@X0 ^ИH6.El*|B:Gx;XVϷ=oz9>0Uo `W7JK{«t8kN*M v{>Z}a6>Mɗ0$<:`S*fd0rRk$8UtWT0˴ s,X=#T qg(qqpQf!ч`!B)i LPe-!b#iJJ "*P THXZ`\BRȳ g7 O2d8PkM~ꖕ:t%nk=}58,-WuI˚ .k:DQGfrq0PМx&`q^sO$}61h9\){O$JjIDs3%JWk a!R|fP4TJFbAC"D)59bHE"if>6|V՟f?{IijPwԅXGYXID`1vqʃьS4Y ' #;0Q8@ ,NyŀЄi"0g"V.ꠌbQ P+rQo'vҊyO;4i"qk(cJr LaTH(̜2zsYIeF{e~}xxw#.jyKh;b5(r9f#'6`YR8Yf,\ lT/|<3: LTN)jc^kOApJ"Iq$P+$Z#g|cSl/w!CЋyRΌڣBIY5Y|uiN$!y rJew6(+Q&8i/`̕B^.HBjZt}Au ?ӹ:msO\ӹ: s$z\j_9 4 x jqe V-䁇FɃ9p;};| z¦|.UL޿!N_񀑢:8+$,#Q $A)sUeys(qCp`~RsF)e٠Sb=$YY'|ﺛv} UK!Jܺ)omn6"gl P,ZZJRbD$|Wnjސ5f/단d#g^y@SF}HnQ`bq#A2A1!.8V2Xٕs-J8o$^5>od7'ogYDfX:B O(RFXp4޽x/57%LmJ}}&sHY%m(kM*u&K*tYQ\6,r%໛Ǜm[_rM[uzu=ܿ%`;w>KбeyyysIפ5˶:hmzxPJ}aU{fj6T!F#tȑ+TbbI$zD`\o:pfDž}Pc -E,THA99bHztDH:#9K.1XHpbv J)Rߩ?FCq8],mGws\g& !b Z.1v:Jt@R(y!`(%(QD8DH|jٚz% `3eeH9D\!ДP.R^0 F՞+ Ty{ςEfN-FIm)Yb"hZ;fS!hqfO3-^^l!O1A`'\A道r>CcY` `,"}<ݿ޸b|hND ,9>{DFL#vȈ GblS"n<x¸T\{oaТg!](m .vd&C'nݞ.- cE!l`*" 8 Mx *H$wсd/.Iޕm{uKޅ27żo2 @ft>5,@}f"w,+Tʱ!d+Y!s(MpIWC_ÿEژ]åVXo>8:'*OԥOm.B?$i"4קO[s}P[\.қJX4/7e].|8ozjj׭>Z33d2Fs't6KS:oxE8ϖuc~1-k,`#1]1s;x3]t6|H GvceV{UvGgwcަǤ@5yE:[^]Ys#7+0Qo2 ?*K"eZ3}CGPKDua[P/y]_,;uD7e5"3C#OPg ).A5((ȯ(Z D-!ȥ#HU6&;dJ~RóO*l L dɢ,1H% a8GV@iu^Oڝ&ٸ}X<T".y[ÅL ^Ǔ'_褷XgR81퟿UwNBD6D.IRUi2OE5KQ!WyCzәU96r]+TSŧׇgnGg01 eNx%AI:0} dѻ6֐Ֆ䕋KbĠ7;030{,*@wa^ȅ1^݅9\}am[ozmcҢwi޽NK'z%zgAvp&g+!uHh%.rP³R%X:Lg70j$oC,ǔeN,h%UՆS!I݀ 2 Bh2s`R+(6-)8GZ:-7ǴT"ȳ'z6mtA@ׇ{>kx:1ѻ3n~hO/׭^^ ❜cHo󭸝ܭ=S37]1% vu{gŵ愐f\P%}al.gDI3FFc3.{]+ `^@|ɔ];5hKt\Nzo1A!Փ1=c47 uY\޾=iP4l| &B2 ⶕ-6.[A!;I&e㖌[ s;Q P=Lҙ\&SgziOs!^ G7CVdlnUY?w^丌s7#\-#f.; ~1]/ɛ"o"oKK?-zM`#ޙ0v-,v`2o^ 7Mհ(˟iA^ǢS<>M&H FƱ6EK>-åON6ٯtm/V_z#eR'8~_;K_Ο0h?(DE;d//d7'vM'.`Ri.}&`يdwe(AL:Зi,&ǔ1 ѢRc"{l8H#ѱo3ٙ;RtkJrR r#wL(5uBrX|!I8/K&5b[,ZJ_,yq90X2n3M)9R6W̓kk51;ru1 z*nz F;*gwI1dr]rQ)Z6Z!j)I(ؘ|kWX?׵oW}S^m {;ݽݳZ0x2w8^JطiOk=:ztęؓǓ!M1 u3-4m?E'j=I@eJ)} 4arKW|[ɘ+>k8dmEk+e xo,i5+A@v:>GhYgfD mVJN=:@L0du_/ʪ {XU[a9;8j"5KR.3lϧE&-h^ҸPVPӭV̷hY{Pb?{?m?OlvˤY! gH0!x̹`}, ljzLYK'BJA:Le|Gpΐ. X;km{C v9jپw }ykiϫk|9' V/{BЧPˮa8U`~~^wIVY,;$[EǸW'* U`2Tz/518B9%!9k9t#A_L.%_ Es F!טE+fagT-V[B}1ԻZiM&EfkdR*[!Ljus6*{8d<8& hS-asBn*hT90]ĤpkJx4uG[^~}em}>ETn+ y4OU{8shpqb.NIԨmY"yIRIQYT{t(۹|wC)T)&)x-QOT9r hl2EFiuU8jRBԻe[NkPqEf1:Qa&() ՆsOr!l:WHm8^QGQM8I_T1tG㓊NFWfB$@A.1qgG\d>+^>ń"ki-pU0j5튚NPG_RpP=کZriUÌۼI29i笸+ʘ۵QbMeUG5ɂ䑁I: u/Zf ]B+%:-hIĪQ[c,e :'\m8a/hӋaXM>jMehqo7R0+CP.Ơ/ςκTzRvH82.JdL\":Ӧ&uo$;d 0 vjn/OjOvqΕkbդvUvovqcD+1 Dt86@L)c_M`*Wyhovqg❵cW{=|VZ"Cыۚ\M[#F_6kN&G>eƹ^ KʖStԅze$ 1Ž3N$2#)7+\tE*)dvrP& ɹsB8f)~gKkK,1_3?F 9y-Z ZtI9%]N. i, H"u!V@3a b$铮I.K0%(+LWr@ց ΚBxBR*@$;1/[8atzqe1@/icN8Z&`hkLִdo4_# ֩k3$% Ān!27Ah SRV)@""=$GT`A;)񱔢_Ty+r"Rdi(zu`Q(uMnһ-idO I󹒔ވbX_%pT0Ɖ,I]~<%oJa7[N,Vu2+ }vB-S@_sţY>O`nFٴ&EJMuaL P92o1 &0-H>D_ TxԊM!EMc^u{#'Y#{6d+%L2&(6ٿrVdʿbs6˗'8 /ɺ|6nG]OƙcI{.WE_#,k|p;FEc{Kǘ@hT>#{GY58G-w`wЩ-drHNonu_QOT)h٫]ꫵ%8c SEO7{pG)e+vєp{\)v8RNj_j]>}_nE:_8Yd`?^A| YA+g/OoG^_5+ϯ^f_l=J-nt^VA PJBأV]zcfحr'Ó'x:4cN5.X c Vx3.iɛnyI>?|Ue =! gr(tVk]hz6QR mf/b>0puÕ=¾t:|aou=f?;ʭHt{wYykodٚI|W\]*`ܑUv']'tGGEw20 k}?B$CQ̡-s?5go\i%RFdC"(4`7cx*iDX i3,2xD68(*\z|RSX+# N#WAJP?i/@5t;H&ƒ^U J+(C;6Q͐jPo|ͩ `\Q ,x@4 Bo <*$@&tZ,E<}n2nMI:!$RF)pȼ2c|pJ$up°aƍ~OE#"ArѡRVH!3XhGWEUFUSY_3Qt?2QtpۊjQլ$SS]>}t9|a:!y's82s`I&̾~6B@Fx4=\L )~:EjC$*rPK ||![R451Eg{HIf nsp1I(D׎n^ |6ٕ;d3f4 `Rɪ%ג6yN~S:$ r"HΨAbDM"~4"$?y ZA QXTDYP"3Ƣ~T}&@T ~h)آȡ( ԚKd/tڳ5MUVP؅75T譈ڠA#ZVH": 悩 y+z`ȔZK uew7vg3r1m?znbI4}AnQ`ơiFh LAGr ncJ`:°(r2HuTtk̲?k>l98բFCo b3 ՠ[- _Z4]m?SP4ծi7t ko}@q_r2QkiETJܼYZc!NN~9+z{CfV§oݹ2_ ᬍZnou}uzSbL昵i».fg'3@u]b9$N?&NG_Dwg0ڳ`ẅ#YA%& X{`AS"<俿!E4lvtWUu6PͤEǗYU^KLp3!VV>ʮEQK~M 4R HEf[ ցb.?gh~S4 ȸ p[֓M ׽A?Rނ½"=rkk|l-_眱eα,#)wTmIޛa˳1"Pom”b^1U(c:A.ٔ^}mF5.^! YA0av:Yw}~5]⅟bN{a}3]l ׊VН.C,~ַ a^G)j!$ 삾dW]gT ~45H3u )h)gTD*iyjĭ19OJK nȼj[f^>>@DfČ P`x*4I;aH|&ǍjQE8pT(׊=¦ c?-ύ]nvdTC ԙbids(X I6c\%[e6rG[dֆQ9V+KӰݾ6yAKeǣ 8.R\{5Mxx{5%(ZoV49]M66hBQ/TlQӘI#R &xfWRMyHw`Z[t->J6 ([fViMFBٲ,Yp:"|5"CoV2vݎB5Y>%V־udXbӘ<_Z; i VRX92€>\ #'ߪgӀʞpg(6볊"[NL$ʘiYb6qNu*UR[xK:NaXM+01h[Pׄ8*QDۮ]H9!{.$(= u,SZ3M-JF,VV2IGk ΏW,˝5l:&Q5*Fg5@P1p1P!By<.9n6sbpNJpS?E uY!#T5i"2+BE!tAÆ#YV ʁ?ִ #e` hjr:H4 <҂&b*"ɳ|`viâ> |erPs}R&+ą ^ FϰVaPr"rkx\{/??X{,?^M c^Obq1J. $Iunl?ޅwPR+=?_75LT6`5 m8E])cxL+(h! \9Nof-eAލ.2}m{pS(P\ejG|=vo_p\((A7L7-R*(~ʔ\G_-uM\q8:\KR(&ՠ s^sUn}~]U 7 6vn%LX Mz#2u `ݱC(ȄAW9XڼOOULiTT.|BwU|F'xzK7V>e<:f)~ "|{E _wkAKY> "WwRiO  r1]lQ(={ZsR9')Z9'eARdso?iCRy,BqgK.q.>sr.${us&ݚM A ZIL$ #ëuL=҅'J( #6ȊLO~tW|N~!N ZNe]V2Ohf&W҃:I͒K<.: gÒ qb,ʛ\{::^}O|sJ3P#Ȉ.hs @W8ԅ8( zfmJ;N [e=OFqR.3oV ˱T]ey+[p}QW+p*i "&H+s1 aHI CT,q%â c{cTnik_D[򖠭Soo{&@պwo(ӓn/*jdžRqnǵI\ZgL)1>Wz%02i㉖/r1!>숷ۥ'>4$UQnr{" j4f!hviE:XYVDsJ>Kl"=`A6*b"AlTS :Ʒ}G[ٌgr ,2GJxڧ?ryS#bCFm0\֪*Gq>:n)!#J/kfCX *QRߪ8ʢv4w~ V4W+|rY3zp~=uA.j=fǝ5L-(OӁ_ qNo6d2vy($_)04.?Dڈ ͤR8b\2bC,,J=rs1dFMw|Q\+&mu)!Mu:XxqB>4NTzYT4oqq1e:Sŧ7-47_97 cb;Eq0ɕ.ڕV-}6n+8?dt2 n}U nL܃#.ްڞSkᔾlmsz=g2YBlՏJ VzmXl/3Kj)` s2Y&<t-$*(96g8jMQ+jR-!bO^Fa)z}yDdR8K<+I^=O*>;T|4w(YkeI!ϥvw9#g&2T:zH2C Kd)=alTeA.HZcL{ VZv6E)CW/G@WŬ`/>í|<@uЇ?/!+2T!Gnm4ӨKcm1dSU ߓb;(`Y"(Gkui":.|` t<ֳ3F{PxOwnF:(ʡ✽2d}2ނǘfIXJƥ4Bymg5` e>VaCfdNzj<^nqDF5hژU+P5(;bQ[CY Nz봊qnD9~~|c. Y \e=˄aҔE#H7DyU>s$;R"KJ3 ˂YfkmH?6z)1HZUg(Pc)RDq~UUWW)59bHE"if9֜rv|ZCnYX. )fZjcFΦ%:CsSf2o^rKY@hN(ˀ=$Z  M&s)h(5)yA-7z;XG]i u8]%6Ze,' ۶1:lv||h[ʬFW%LX/Uxf[% 7,0R9.x^)$19Hc-8VHhF8m/cJ7td[!J,[4.ficufj_)"e]%Dr=izg9pU~?[?TlX吹 n'%і,jWZX x=KUv10Г9@.KJPJjK~hBLMEn#kfIxA'B<2].h)D_),S;]9{! RrחࣇaNC,bd )Ni@K8TJSؠ.Fd3UH*:5vwwpY7f6ej샆ɪm]{ɼ {gFnml.Be-'&UZS*-hw1a+PUSrYz{0<쇱u4 d %zvv~Qm+Czvy0v1{֢MSTηSԻxK]]Aug-Ռj4=i{)+q%b# -JR).] mtWvlQ_ϩ||E}U ipRa/@#eʈJi 18 L( Q`x-xؑ:>M)0pZFe)A@gLسWs5#\ick_.+s^QED0WԙX i5JK2pDXA8pvZc -AThªq(]*ǝ6ٚRp9^XJX#PF!PLA5y) >qb7 sw'KHE&SZ@,= Ǽ#o ZhO[ Hau/ KKe<oX(9u?gLt[3a@9%{5ygRN}O<"mP% S WAs;L(RFXp4VLw+_y<W E _T\Q1ri˼=(~}Mf/b4TS?w׳m݃ om_@HݐLw0d[MɝR;ƷIO[wgq-|T@xc;|4's7M ;y|k1u%%ޢ~v? j-4ZDIHSTJE$1똢 y@+3Dx(t~{Ӫzz⸐§vA`'\AAr>",z# ab{lAu` Ӻ?8b< bGɁ5dzH:h4DPȝQ@#=C<'O7^I^%=".hٮ#pH`F}(6){72^ʡēׅy^W5U %( Ak\$889# ED^oy/Nw`OYo<=`ϋ5E8té SKJ&+rd9(u黓muy;p:*\ZFoU!DKl/4ZB>x+ލib$图_&G?hyTs韓 /K3y>.~LN^S.@ \&89iR'HIUIianz>XtCС4ZjʪJ?n%ŔcJ@zd\2e,I"e0~ H ?46o0v[݆ѫ:&6%F,EP_HaO>ɱѺR5" P ڀ/(Q>?,]g FY-寤(M5{ xzvr1ſow}_*|6E?Wf)fQk0Ú#"i cdqםe|f1brL0Ux8p7k.9<Cj^h^\~&桊AU wy]'GD@uUЏŴs۝Ax iMwQc,7am\hlFW>_zj]0Gkjf?Ǔ@v:,?e^_ޖח ٞ$2j=7.$~7{,tpUku=7jN9;^Ue|7kLlLL40477U|k&Z3TF]-OP`' iV 2Ay1x"IfTb@"PqoAr0;2'<2ep]qydabNE`4t!JA pa}GY9 Xtf[;{Ya:W'#.:_<|+h.)Rvw’h{ s #6b>.]33vq5ΌR+޿.ކugSu1(~`F00+v]u!1t2jB53=?{~cLN >XC+g"Ke+foT SÜv]ISSn[|S1r8O5n!lN T,'ʙ( (%kuD;h[ 4ٴ(Qw?cgmI /tJÃ&GMKpfh~,KLY~,w5h-#I0/?qYK֯n3wjLMɱ#*)dFRthw̲˔j՜b S UXWډWVKookvpC8FWľ]_9(GhW|:+; GwI OQyD^Z bl|i4H,k ,آx"iؖTU,Fe"39V۹ )ee+5UpTӃX[ͬfG|4[M8ɇHc6 >2Su'0z1蚇1ht Mn =C[ٟhYYqfn5ɂ74̺4N+gy QfX8i]?S'C:9y[>aF{9D/LqpEHniI)&2;ƛ`i II\\yRjRX$WJ6`ҍNkɞ./LG@7m#u*r3/U5x*k M/O\v^ /'rrv'NQp9ltDB$sujzƿ5ZtDmiIUʄ6U4pbLYH*[(kF3֌֙FT?munl(nಖd(:FI .2T~8*K'q:)o_."QGg˱G7&]ȱd):)E3:t=O/T+՛]J3e6},1{(lĤLV+QQd.d)iE4Yv3w9CkY8h ~$-EAw =ⴞL>|OwHWϗҾh;?tc3Fpkw+}ػK-u6zzókvڢ]c렵SQZy5;qgsvl$7ꌝF*ve5kjޘlpL&f$d u}@IBU28F)8i\c,'Dk"*ceiߋ#d&7)`KF94$) )88PjC9{.jʗoꕼB貪 k,Acuz{Y?oF LM5PL*3?c%=늗xt^jᎧoNuvz]u"iu~v XMY'm _>$rPsTlGgk53ɏΝki-I){ l~~zW(:w%,Uwk7GƽQ'>$Lm`#^fi*s妡c;;`D K`xd.ud9Q\`J)XP(~1}_;xw%ex!a ?1Ei9o8/oLQqXHf(mI_{wbZc+g&~+2x=s,mzMOB*eH$p>)C.ziN@PJLy2$ZL(1GBL2X!xmRNHp›$ɱ-CِS7d+tR J#dpL$5>k  (*|Ʌ$% plk^c)43jiR34*#2 #(m ݸtJm&{>čN5rgj.[6ecۘh8\.)Eh+yX4rt2&-P/z%JW,mllV_@kl=sHʮ'zVHT4>M?LW5 [d_Q/6-'>2ڠKu>ڹEj7Qڃ:f>/жzPåyg^<:׃ J H%1]W"-EMNU=4 r #~27??;E?v` & [ns9_d֠@p}lH땐b!/<tmft}kKO׃.yEY-<VaZXgGF" GGH:pAS[{qQQ{4.8W-(OӵVYftd@~|(Ֆ+<}aWT͊5Z1f|9f>75iZi &2c"DZ9 IV a|PDnf3:r;sM΃{-:.!(lFύBE-=U̷MbҖ8-Qp0U}}lrEN$O2KEY> ?i^)Q֮oT  r-UoDQAa8&_gBؾ%OB,'B801d 2GZGPā*ho)uM I!8\d3KgŠ;EpXnx6mM ']aөaAiÅ[v;qҾR&Ĩ&JN&OfB xA. M3\dRUI+^~KE!!*cmז8 /R֡ ss׀VVmKٚ<('cRo>|blsg,hû"l֥{ztrܐӧVrzn͝`4SoCEVLg.`D 4 %I0M |АpRd'uYCp:G,v[[aO )xִcW[赏)wlKƭ S9Fb+$u%Eײ_GxOh!4HKgsY3(@+l ɭr-/zg[ڱ/_ͻ+?~hWM\x\ٔQL_8,M[=ο" L҅{Ia{kR(L)0J М[u4SB T(?\OMmkniC~tzucDSNBq  118愃`EC1YSh,ko]A;/Y]uG2%wqd d7jң94G7+.KU=j43M&5Yߚޕq$2{xK@8Ļ@>%P! ݷzx9E %(TUld`z1sW~l\풁Iz5g_jX} 4$HH ֏t4 kFa52bՠQyI>=woZ捣2y"FmZ\\S(G'#_Ȗ#q~mKX_H[OWϫmo4:ym}-KPW(q0J{4:DY+ ۰Ttjq{g${w/_|w|27/޽9sGsM7Mý&d_~Cx󡹆-z[%7ݍq ap<"D# 4ݕ̷Pw'%+1ru[s7IpAşă4Y%X(ZBR9F A@.!wzC¥C4:xP1!h!P#m[ndP4OD "HF|@|t38E;91Q߉I8-c۝Y\*7XtY:םqADBuםPY~Ɨ#sՂÎwĒJ@Q:Lr*%,e{RƻcJ{"KYr >D4$,*>Qփ$ XdpVĒG4PH !Q+Ja劄L_gg\dT&<{ܩJ˛Htb>TiWV?n^-xN2W| Dp@J * *]Ux5}\1pH<&k2ɅMAKa%Ss`頂&Ć g^%RhDwFۗ)ޣ5X?\#/ ifg?)^OjȜP" \$ϐ}Th2~~HYzOOE&C;-3\Q9KF:(yE}[2<5hOI1ۖiEnbMHS"eL]0ok}2i 81,#N2k6;ίq1Fo>zmGVfo{DI9%%-'{py &yd2xZ| lڽ?wOԍOBlu{1ד89)/lV6w{}z)kͥXuq;-.sw[rofn7m'=_7,&qEK-7C..)kކJTžoxVIvӷXiΗ?onM!-py~:ࣤmE5𑬁rZc9u~={[n>`EED`%fsi:~̂;UWzHjc|m,DSJzgb-~b$\Gʭ <%OMTnd T V;)፧:c\k\y+`$SښF\ `-|\Z W̥7Րu7^U:L5a*Þ>^ l"Ӈ ]5's䑜M4>^Һj|YyҠ}[R JJ_Ju5f׳RHܕ/u4NyK/i}^B״> i}^D3* IBKtF@oH4"O?GCwVd0J(=0"AL(>ă2Llo i=t!id1[PP/}QG4 >AER h EP:bwOo\1>AK%ǓĤu&H'VF" oCzd|qq}k\mXYWEn? 6;5}2O%Ԇ 9ݴ6g.5JmuRKNbPU$ qD M*צ DEKɣIT'.wIsZ.y%TkoQj {۟?c~o ;U̗/P^Eg9I1o;_'p_ױYWWR&^(3:cVx2,1 1HCR{0sLH/ѤӟPB4|TO㿊?uq}oca;~\WٽxنP.?*` _amVX_g$h^FRL_q_5.[>kcfj4CYT02*f#mαןNS]Wl>ZC;:i0gȴ:KR?Ӗ7[#|q4a!M9AyN 3Bz)Te 39VnU[\m bػIY=O vcafn'olw{S)3lcYMD.io?nhn XwqgiLdȓh2Yv<}A؂d|:\DPy),jfir%ZQ!H ΂E<ҵ_ _9O"^P1YHΡD{KCAyFH*•1(tlGe <PNu6V_gcN~Zą/8==~Hoj?/΃<烪͟bSd, o>v$4V$KSr_}?Ǫ!+7ԞdߗˇVsl"(˛Q+]Xܷ}hznڵߵAh8HKU60)yH9&{:1hEPZTD?9Q)4lNEG&s%F)b_`R^CLy:Tޗص֡%ȎC-=d1s4@^|%XPdm޹ҲW-Bzp.[mIOyq*(&}7,_#XЇ "J%-*U!cr"cqnn41n>/k_Rq-НZJʵ@5M)SBLµ,=_9W&g4vc7Ccv>alla{-PtUsYP=„;B}s<)%0ܼOϛ&Uj75sn㮢FS dpmm{$ד.(1Gpר#).9i19d[rcF6o3&s&?:ǐܒ.qFcOkh]dl#rԌ1$M^mBl1K(Ë߰T7 ]0V3{M6TZtGN0Q9^ E*9c#rIc)7Ra]**5(:ݠ-B Z q9xCyGtX9]%Jz<1$kb5wp&eX$YCqah Y\@Wg#5ߐm  6g,(PYjXౣCc6 хQkn rhWe3ܳa3bX؎qBނqw* +7I22)"=Mh(pu6LGqIKd|FEFt(Mq*V[!zD冝dYcݦ]Vۖ*dWnltEWBmTzCɸAMACXg뭋XfzIc-AʂA J@"M=cX AgE"E ÄQ15K$" 05 i8tRp 3G]7<2Ե~f?V!zoѦb8S %n6FAu!BEGT@NRl m ~J x]RY u/ JIW NT0+-424R( g%㠅fIc3d5> Dbѷ2l-xZck!z_4y^Ƞό>tXZTUb)8Yh1f(Ԃ!NZBNH2m9;3l7xaŇg'xuN B <w[_Qv`5M Y ǀ 1rۻXo٪7a7gP!>,w EkMjׄ>Hg9B|G"o /V1^8R.1-b FZ*ь$<W U4~ ;lœyLE[c@s6[ˈܢh^;=֢U 5j/PxRQ-Lb2x|UU{OH׶G[:T쵛X;I/{u4׷˛?ͭՓ;=^>AtAgg;8:81 Gӆ//ˏzr)J??VW V(qUЯCgj7oT?u]I}F8LFn-?}:⿬_jߌ0޿~M//_9o_ _dpU&fPi„:BpU 2kKG 4nJ$˗`!x "h1vp`z1ƮL}>kI۽4_dٳz]jVxUTFT#v1,[iivx2="%ջ)޼[ vĈճr;x]nm#G>&s>>~[>no9bק[r3mYW޻m`>腐]Y7AKotw m1`'e:qW8.gJ82V߇$/@EAc^ᐒ^2DA鮪Uur]9/>`"KV Ä ]lHz/g@Weͻ<+hAEi#l[[:Ihnl/};̇u/ dO1rh\+Wg֘kK7 /pkA(iEәmb6>|| eY\ۘGlٿ ?_nBhl|~e*PX/UZd=MkhձC[A"~s7lXZ>S'ՏOl: ]U^SؠR& 6 Ϊ|r=5>MdQC,::f8526+3v\^3-ބx[i{oLh,}T1[[z|WnC 'pcfu4ra-e?0( L/FQ HCPVQOluiN(7ms>]DM:OJ! 6>ūϼ3pE l %sr&8bR&MnنDgtˏTgϦ~y7XIW( '0n4,`jwM:^>^J'>q/ SrIYPBi) 5RJؔKt'0>q^M]r.ȱ֌GS-qqgP.(4$]rf JIU"bד QHǒQf~DdILy[Z"i_1z ]{av൞nMgk3..?;̺vjQ]5]\ (t艔!kSb%i():C:íNբl[]qpg!.sQXأIԒ? gK,/k a!y|#XF>6?ωqlAp)xRb l%*MKtJ CAkyبJqmuF0w{{_AhCe }<ޏ٦'0֥*h C$%hJ,lB2ER`ǰl-"04Z e#cgf>0.S"WSH&g a "6eՋc|c˸slr,w!>G#h Ka Ф,dy\{RqpHMDDvg̀ w=|1Q($^ + N)$dZtr|T^꼽9?;iuF^,`VK#q@>p R@|,(".ܦW ț+Gnڌ7pJZOq) =3uo-aɷ{{Nߩ q*S~5c 9I#Qjtw's(qCp`~sن)Ӡsb#$waؿhF{ԼmK-`S.ʏu?xw뤛:Kt[b?l5DB Rb  -z YSoR@}ՙtcxb+ρ`ʨ- T2QTL0n4HV3 uyprS}lҩ|kY/2[]I^ɟN) uD{˅.x<HTe[E] SjM!4fa-*TVl)`0jq:v37D Jr< VoP2㺸uE7ł9 ˭."I2Ny#['ߓ㵠`g7rj%vϑE?g˫/Os*%'&VQgh՞<#`HS]eZ+Ioj͚КmJ˃KT)n+)oC/zw7aTN[ >T>L/zͻa+7o-}6/GGQST$íA>pO&ly¢Q.o{%`;gW[׳XWv%IYkRكi[v{&4ͧrêֈMU:T!F#tȑ+PbbgZs{ǁ0=-{g?O1.`9B GQES#B9^rB:;SU9U[YE1q2[ۍFVmή_jCR\b4u#t$P4Bz,QJP:phܲ5us 82A9f,r`BДPR^0 F՞+ EcY `,"|< _;b|nD g=G"A#@^;dDA #wGG)7sޥx@9GCİh-H43{(IG wy^@*6::D%|x"ӕ=~HopfpytMGgY8hB>)5<5]SJs>W} T?Xw/H !,:3kq\%`y?suOmM:.{L,c05%aJĨ6xΨjgA%GNgNA7RAV`j0tTgRa"< kT=hEJBG(PKXx:Y:9|uFuT:zy׌GS-Jqy0{D(J&%4 l@Q$ӌu0!(pI*faveZFg5,wM\9P{XJӳ9d\)V>+s^QED:uY i54SQ5b91Ja1*00pKM:DBX1JWq'Jdfly(T+\os,g,TS@ms@A[sC1ee\PO,l&XA:f9ޞc'4dHE&A ,+ aF*Z>eۋ>m "_`Eh;K(OM= aAPy0tg2)06㑃xTūq}K7"y|6x9vUb!}2-Э+H3o A#.\*!()VDPj2_u޵$e`QR`^[ nE?e)GR艹!)JcL5-Zp8Ù$h/#a~?Qvٛ$v6֒2g+Q06Dŕfh9˱o'-dr0.ǟw-J]ľ_yKS9 $ܩS տ?^+k0D(]3m2\W54@YܣVCDWsU?P g~VlgYsV P5B/hf"ˀ;?/?_]Qޓ*r^v.-dV ozUyQN8.wE;c?$̵`,o'˙يDKO?T~iUpUU1U ~hG9JUQo/:}i} {m0Aߣksw96H\lS @k=CU? vV }{9eоw|qΖ Q0<Q٬/_mW ͫFrIl@]zuԕ`LP_QOqx=w9],N F?FY) 7j;sY (ךY + X3 Cz6 vOZXˉkLAѬ'm=և,)pYeǝ)HRZr#>[7]$Q@`SZ3DZmJXvlqBrr0\4WӖ@5 #kCL~oQRȗh,?\q?)l-/P$=WZ<7(IA4ժ7@y]DbB_[{KwZAXۥĤ>>Bܯj5TwJD%z迓H]ud9]#v&/H1nqd z~ JQ@Ι歜{tLgJGÎ-;PB^c&~&WBX 28fsq@94 L޺Q$8P9ȥU+P(I!FzNG4HR-Yl8:R(PL@10Z& Z*DL 0J-!dBtbsQy^ůs;;}_58fPվ:y֡kUtZh4ǾVo+o,V* qExԆjg\gy$ggٻgS\g\WEyҷ]b^⯛.ӻW8aN~ŻДv'ouUPBJ36f~-͡jqD Kle'Lsj!,XcU,fsMv9Sih2!֗|t=ݯu|2<}griJ4\;ps82e^bzyEw1fIAyУ-8z 7&K4^mo:D{ՋSϏ-e㣌 f -ONGg\ہfŬk<8l_Pn 9B@ @黎=oK#Pk_\tҒEկ6./]~B{@ȝ|vOYeiZ<,jcv{QG-'kJݧ>N&9Kօ@$dkU17@S/~z7z ~.{{|+vd=*:_:GġkVdյI3b(֫c͗o\ǔ"hhTR- ՜>H,}@IǎHǎ<NiOJ3װ2)cy25%۽Ţ,.wUPB\R{\YRm5 c]ԶvhrSq5/b+zF=/jo(4YEnMURn;8Bc &KXP DZS 1PKw1AxGiWxWp^15*bn֒OµWWF#tOU*Vdy#=z$v57;z&v83`cvwf%Čԡzݞ -O[b,ku๾v##h;t/r.gpqsA+-y0M}TNq{o9Tm+C59mWyX1[6+cqI?5mc /GYuqj7g\O.?0Es9x?^'[ԅPaÿ^JǷݒ @>vl1bIL+_g,dS^SVJuZ+\r9OYeNfI`˙"c#5 ԊD#܆ T>?eiZr8a0 lɘ8=YjNz641hbиKJIpPٔ25dDFTXg3@mƠq!(\Xύ$!\}1 xVxF?zj*DBҠb@5&r>T ãI3GP_Pvh(FVG#$K3J1&*ybK+5g=kӦt屯V%CMn)r_q,ɳT? Cc篅4C% 8YʈI5KܢW91ZY+4"Cm:C)8{CD.q *Yw.hdD$f.jTU@*%QY$(ǤZ*FﮉvB}:g(@,JlBȭƞLG Cq/5p?~OEDH(KFrQÃtD"XM!R.UeSQtT$#4eD1.@)E/]pv+j4ICT?uhc/&'R HrK< 7Yݫ>aCk86Y !>X8Q&gMH~k%v\)o ^5ǿEQCEN<22%e\{ǴvR((H]fhڡhqY(RBr:uH2BdC;!TEupv#c9R IƾX( c!XQ~uflvWNoOtPU^XS. Y џB-Feql!ywqaI.DD02Wt[Fl7XPtڲ0jj67rB52@UN{Mʶ 讁ɅX >WF($)IHeӂz )yT5! #^"q.$)=:Q g76N}=釘X|싈0""qk{J„tY7h$&di]?-ϯÿ,vxO2Y?rAhba;qVn1ID$^ޏC>)+tl}|sկϫ[ #͞KS]nx޵q\eQ\@ ЅI] . <%|($%Y~YrIJJf 9ǜqV=ȏ~ Ԁ1N(+r|甇c7OǩoF.KxB@Ik$[ 7DWQN*T-f/#SFr>UWQ$Y}JJyzz2\㩬 VZ-=ՒpxqK*/Tr>=Ӽu5k5qSuvMs׫l`df1~r6.vAKmLx}醋+AȦ;INB`mh6-bLQ`qq' wDO zmgƻ2Cng5J=27|~·_BO =h_60{Lwm zKVVx9궽I]Ӯ0*ؖGUR6{pv K_|/oο9e>{G^) ?ʦ!p!I z^=xk_{knV>|]>ྖ00 1!6J؟'%mvlA*b$=#F E+K;A$5QY;#W`;r!:P"hFDALT;mǭ ZL_fI좌_NGqI^_L{WkQILcOOb26O΁,sx<4U%ug;6Jsa )pcVH )p!.=2BqU *k`BV־ cA SVJ=BV Z!X+k`BV Z,mY:BV Z!X+k`RtBV Z!X+k`] Z!X+kE#BV Zы`BV Z!Xd?rI@!X+k`BV Z!X+k`A*k`BV Z!X+k`b'BV Z!X+k%ugG䆳IYYy'(&eZxmjc nx5c&d {P K\ ԃ3>=k gZuѻIJT4z3t? t{rۿ6k::fr\}48 1RXJI]< k2)\ "Q5,HB0↩܀atZ كOz0Ïa̸N;U0GmPߛh =Yd7o`Bl,^qʂf"O9?ߙX˂i>ɯC-ɿ\(<ǥYGa]МlXҖ!YsEy RsO RG"+?V8_$93{KJ>;6HC["g֦Ѱ5XaHB"Nj5 1y c2|E|r,FgSevb/9Gաf抯\ƴJʵF)a+J+is L ֈّs >E Lƣ^l]) Qeh<8)4Izb;rd"S ܵO^ )LxawZ7H?ݛ+w} iI JEKJ崦C<3GYuZVj֭S%Q7@u4sP୍c!TĜ7kf~ee.H\ V#v1&9{/J5YQerXgJ3pl( ͎+mzwv!c~٭A˓F1[כÍ5\h[ew\s-\@[+=f[VH1!\WNw\pqr.øf2Z{ΌT%" z>G1R ju'~|:D؉ja1?U*E+VQ@_rsrf*ft[>;f#5Hu錧"jUzetG@@r)+IQu3nEMVՕq㾀Mѣβ`!Wۣ WU|xSIx>ǎȭMF8]AXk)-1FR~_;>uX͜2v(c*Z. /I2HE$V@1LX1Bq=AXmμ:((I%T$ΓxKeL$,KRFjsnD: u' :dL|Bo|7[s?)$S:LI6fayr` s*5:*ۙD/O.G HYZ9޲g0yR(x=/ZoIrHv\3@.)QΔ&, fIZ[U,K#&% '嬳)gW <Ԗ (Қ`+kuY{'e,'۶9 |'J@F@2;ذvw'"|- XK*ʤ6s9SRВs%(JRC {TN<`CVcCu,kt^Opc u;wMI=o%T5of %3%+<$'h~Sfbo;H; liʾنړ$8}^οd^aKy CX;d鴣Ik1h(KDŘRw mYG]֥HG DY  Sk yJYж솶{ R- &5y[=%2>(3ALCu.WO;H f &u- ÑXTdQBO/E0"C9vرt/r/TNd:̲DPIdT4e ʹ V[*hHozN5f%C0xCmJ ,zpg($F3&QY&t /;N:)ry+>r )*#(e1f2|y%O~\!k4wT°{CtUotAZ`{]`At9i.ko2UǗoܜ opTdo..s\ѵHVyQ $&c:F96Mz`Ə&鬚OKti0VKKbQQLE>HP. kz]($y9k8MgE0#,G֢fo2+{_!ݬrFI0z(Aa>h?ʲ&%fXXji{{5=޻tVdw˝-vpgPH4"fp:kh7=vcyhx&^{tX72&4@;̰qInPY{v"^;ȭC&R_94kw" кczz{0ɎNmvgUmͦChYCzZv~y0^[>>k~ZtGdznN41<(bi67۶ KzVo xn~iL6}zW}7L6jҷSց/eҢKFk-XB➫9l_WP ,s8:z%&Q۱/I4.Wч 'ɹzCד߬e*?~dtow"c5J )x%dIQV 4(y) S!ZdBBLXAV1f)͙AB*l@F<՞vԅ5:n!^9!ۡ1u.6OfLx+ux4H7bݯ2rezfi,b~ە*/K}7?+\.j%":D.`m+m#Ii#"`,vڽ3 x}>!eEhEUyDD~ ԊWrz OU <*[ XVj8,iơS㑉̻n}@t@#d4[9/ôŶ!h8H>n0IC21)zxaa"mr1^gա 9~ؒ8w\LmȍŅ9 C0܏ztBKe!jjpl Y;-iP>2c _uYoZYlIޚR,MJi 2&9”SѳՒ$ɒfȂy.4JQ#e ƃjǹ2D&…!)Zm5gCanMTN|16ف x4LĸBsbİd9a7QQnEVD87Z[B!E)N&.%:hW%֚43\}QN) 7R ᢕ$gqPI ay>D!'R~dEY,MͧfKv1 93yfRK$1$mt"GF.X&B+e򶢼S%ջ i5݋<]rWm .z6إ̃mohB81ęg7]^k-@ L}e HR gsSQGgGG*ozOkl:nN!F J=᠒RxFLQSBLx`O'N .׃8 -b V SxiFz|˻>P_(U$D)#Ȃe掭Q8:p'5b | iAI͒.p0B$R|y|X_(ud䩙~\ǃcvEqk˧AXbXbsBQ y' gPB RuYWP @ tA;PJ#4Ӽ} :H) *XL[3xoWkSfH޿ڠ2?nM+%׷(WJ|=A52)n %9%pU)AAs  pQӎn_bz֚by8_Hr}nhK> IxH#D 4(s%dz(7Y/Wm#&hpr'&30D`<2ʨIxԣ'{ /Lžp3jZrώ"Q i28Tdܦ%J$.g[YR^<Wԗ> m.>msE[ PNPީb|*:I9e;B_>^Q6|^ZwwWz1 ZSC ̮o0/&1NYe+^I̹ם͕osz_o!نP}&8̋p'sfIZb}[>yG2j6GyD;PH=RWQgu=Hb$M7m)3lF79fj./$`~6Gkdǯ~ 3v6]pb0ysP;];w#L}u&L]ѮK|`8z3 7ISYA5=a3>1/ ㉘Lɷshƾߐze0B"7L<yŞ8d|:\DPy),jfir%Z!H Aje#`ܱ'eۓǓIր$n~QTcnn>sU/bV?^q=Wu=-- &//|e~:C:ۄw|I od7PJPelopVM8NSEݥl7):bC 9D fݮ_ ۪cLNqi:XA|_t2'뷚d͠LV/;Mcg6Hp fݦ_@: Ԇ:tNqJb9'em%\=4 QpI&{u9DOLZ8lKJfĩSr6j2%g7A+g6g3 V=`Rb6/!ʦ}6b*'dcs 8gB}k:e58vH`+S(V,KI~d]N&>d3)8fp8) ELC5qR Yz6u|.k* lRӂ nJك^]7P7*66!xgξ=0݋gHMt;΁I,I+0|!(e"A$}c3 `KC8FߜUYx|6 z~ÀpMn o-P B[(x 1f`POQ*P B[(x o!*LP b o-P B[(x oTSѳՒ$ɒfȂD:υfB)jcŁଶq. psi h9Zs6ִ3(bBRnF#?)Ʊ]<+KGo4$>hqO9Ĩaɀs_o@' G5NEӚzx%RP\RHt4q.AAk/ .iٲ֜m:/e锲p *.ZIq ZN9P}@=Arpc%+reA?4', /^idHE唀rkR`*MP`(CQHxkgiR~3Bx.&DiZk׎i@OшiBjH/*?8 \a@UU ͫv Ia0<* Xb=He#AߞJ!vG&ѡ]R >KS03o~[cvx.0F3't&5 6:#sQEy[Q)z4^\[`Wm .>KEb2ſ9)P0oQ.rE-g@++"@A#h88,ygvpPHȅ( K*EMYtzBR$!7BvwY"ur)\ІH;~Θ e?=z &CR)dKd";ɬ9bfخܺ"gAabL᮸Z[w^k:wlKƭ Q9Fbdw`CL R]f6naV@2!CdEGtMBF$ 5GQ)MɨF㻎 wFf}}E1F?M5^#x`Vƨ\JQPxEǝuXz0;CgO82.IeLܩFt3 m $wY Hz&-h `7mu;٢GH/6LvW/Ž^\B&y$09,ƟQg :%.Q238VE [:L/+}x2H(򱚫f"/9 ?Ipg~T\;8mI4]5'' HCv?):(끴.G_@srDv$%|ɅK]` BY**A!9w@i,7'9B[>Mص@?<=Zn5MZ8+R{%!%"(1BU$(b$7xިcҔ;$R(LI0J Ӗ9SH8H52]#//u &*5[leڕ!<&N{*.HH~뻓Ow?_ߞӻ.ɛ}s폤sc,W $A~M7l?S|󩥅-֝>l@bޟ똀륎U!0XO% %qRvZicvN)pG< Nq遵HƘrF@x\& v8̕KJϿvڝ*-$R!NQrWW"fT҂Dj(xr3]ĽnǨdl_V-ѝYWy˳VSw'u\YـG+((@!PC0Z::)Gx2׊=R}{M$+};Lt@ȽdL3yY*d %Jz%!nO;Fip,R-|8i\c,'Dk⑖30 Z (]W튜> 2Ԥ \j`R+Kd{Р}%lyUoΪ?Mh]`:Vg;TzpwMSo~0wUs1g{;Ұj/xкr~1ww:~v_uPrpPO{1Ս jTUjq4_U{xim3}I_vtfY^oU/!r6BBqۛpz_N{xԣWߘɚha͟o,Tom%FR θfnZAdLV|SΟh1 ͅ^i  B`8Je)G[>7G+W9Y`eV]2JW5;s/|KZ/GuKo=͓J$p>)C.ziԹ)EyGed`2$?W8߸R!c%YcȾd;[H.IvfI<<ލ;2u}V ,T[v53q3 '%h+2hrB# kр1iRa+!Њ,eldlܭ=u7ܖhX-{ݳLyt/ )o#wA/a1u`5S@+=7}{r'zV%sTL2ʁgF&[00JYHRQ:O-+Je:C* yZЗ d{E:#g3$ ]_I[VR]!O^Ҍa A[Ps CgNG|?^2di!Uc LIjk *4!oK.w܍N{\i F UY4y.8xQ9VЦQ'0 We&-XTK%ǩkY2î޷G'qycT%bG"Zc迉^t7MPoͩhraPd8 {dVm9^=-ZFPn25 RptViuGLEx y-B%}D%1feԒFdWb["n]O^?<-0 Uzajҝ59԰o"9Ɩ bdЕd%0ꖺ.o=婲J:G.JͽqWrRT,D EtS4&x^R"xbSaȷkTSŤP-И=h97?=9]<]tv~z>jbrjSꀂ@"?z/:asux~>Lf~k}oF9w[vKܩ\r;I) K[/<,9Pm&d'Gi\IGm7Q qT VobJ8P>,UN%5Cè!ukO5s6Vϰ+Du@l,snFC$NyJ$jG3QKӫORuhrbt Ix0?B Ŝ5ZX;-rw/߉"xST9e2";=Szv#Ru8w"B1ht\SqdgŌBi.O&BѸ.c̆>I#)&4sng!p_k/?AsPJ&s̥ljyr`#lz"Ɠ%'nV w_%[/4#QgAP 6vlsZ Z=]uA-y|œ_Q^3_ H:EYRWFdYRcDJSǃ?'>2M6a X:jabFY6*vcpnMc++V؆9qG~N~(7E|~r] ^l$)Q2^<qM&dTk G]1/<.؅!D4ktMP=?|[րo3|m@BOHXMa@0ta}O]v/[ Yv!M8F >DUKu.w UjTNM}jwj:=SA6k ޛl7R MϞ;mj!UjP]d? 5L°.WycnS3ALRF |sjyu=q_r"÷ Y򻓏Օ?yU4PU[(ww=8LDQîZԉLxH45m[>> 2ڎe  )1RA-KPjrO3 @]gZq Sd8hɻ ɻ~vQ7zN!gٳ] wu}?t1qJ`cϣN=knw'䠍з|G|/,1gp O\2O?~ֺ ?Y~K[ 2%ag9ݓ'wg e7##3^]go+nߋzXei^G(p@⺞x~W~z'p~_=zYg}uIѿѡ}v膥>}߱OGCϫsS~˓n)>~۾MaMTsѯtK6~ \wP.'m|[>IЙ(D c.=y=&ɘg<߽~s22ʶǟKM.B {z~P"nɉ\^`Dom}۝SzV^yHG9*wGCg ㊪OeW>mZ#@"yo aòlt$yO&ksuKuVn湡wWz7jjqI itOj 5s S?B*6-U]wqY˯g1?}d O \?O_u1xkw|or n+d;BӨvs45ODk 84+m>'ܥuOjlS'r>\LiсƱڜ4rsݛM EŶ.[sb=+e=o|XD9#oO|i UG!oa2S:w˥3rFLµ^Z/dB 8  9*nVزZ Š߻ 7x3sf~mmphx"= muo5xqZn2Ao0!thDHqHp ;PHs"{pêwL/ʫ_ϋ/٤X'pCwMWd\V:I± VJl4)}D-(%Rl6r2#$m-XIÄVif>lzSl%쥤GYGw(k" nSAc~3&B_i,KS뢐%6-dk %;J2(, rK!dn&itL]Ykp(!6-(jvFl)U (V1U&!32Lc9VgEDg;Ak!N0u(@N6Hq6Aµ7_KKH2’W26bkAꖱ=d qai=#^n4gmUO'`,fE#\ ZM`)5h'Vz TVz{UQ\Fq"[")g;N[5ywuTX{ˆ 4lB;"Y!dRߍ-vPtJ*Pf0Q ˫Mc-%Ds=܀2y6FmTzJ]jC:ʥ^lR i}WE S ˄1)wS`Yb1H&:󜸅XͶ*Aέ`@Xt*${sK J8 92jܦ`5^N]Y6;KQ@hկZGjK {v*Jq&m.Ob72ϐ\i,Lcgt5XkR>aݙXᗦ;f3)c,HQ¹G# Nւ!'d_:ü =վ<#9W σBD="?z]#& l3RFB$n{(yi;>-XȦ%Q餄6R!$4i*3`+i<)Ԅm _O+nkg7* ?k)bH5=Sc)` ]I1u nz=| %[2bF_P6lE]q("r܆ez$rN%&o/o`\8q9Ea$qdoWUgl5p?<)1IcǬ +80'E1@$jxH5dm2*ho]}fң-&t02UE(5shåt+7 >ݕI{|$k6P<5Z3%aej3^\Y>2"#M v*!SaϖVrU.MYMDiBXMD%IFRCl.P,]pM &N@}W_ກ C@︥j)CgG 'Yzz(qwhDZ>GbK?O28vv BZ%GpF 4 n1uŴ@}9$>ݵ'S9I{ꆹvfR/)@#y'2\p S QPP(>&j';{yz.5hNFf\Mk#>mjugtQÿ\x:=X&+in q݀zxy`q o)_x@l& ~Gf&df_gwlDq|MmVm(<U;.U'F,:c,}2H[äg>uKԾ_STz=p3Vܜװ1e;t3%n'ZfX ڡ#>"XGmy5l[[{|֛;/oph,4}p0@Iĩ'r+[O!V: ,9פå̭0 gqw^#iq5'#Pɥz4wT3uz^kU磭Xe^TK Y9KYn4lfmw{֝$Fl<@1]Dxt`%xV٨tME!_vUJu]eN̋`NЖۺxkr/8uo)`bǝAg m`$Ө|- m #@VR8#ŪDꝫb`+32 2Cޡtk&:T[o[CYڵ>h*5mONMMsms˸nޝ&hw.BfnllOE? ص+FΘ0#D|xT7߆Nn \mo[M[ H egN\v+PcjoKuF7p㺶'?1_5}ܶ}.]Y:G@KףK^PǺ$  HGWJf P/sZ*˰ '̜\vEOmo[Kb2Fʱ6ޛUz pXg9u!S])y$QE*Bl"BlM C2߽L`> ;o'BnvUU=t3G+].LO TyzNIɔT4agTDp$5TI˓LTKdeӑ?xO->W`Q&as~ю쉎$bU2uD鬪S7{ޙY@KBNm7Ƨh ~E,dz4Ÿ=˫.>S8vh\ N9>}.iقCتdwU&5&Vja2P~$͡8{ҾCds)x Q"+ĭ-@U[&i' F G8!E8ښ8;:M¦R΍`N3z@ZXvu'-(X&Q h*9.ׯM%·4,}56hU'ѓddU2VA1&݀Vo4|d6%hZ>5NQsɽ-%a_OU{ua+?Vmi$&xfWRMy- Zjeؚ~kuQg-&1HQIpWxdxtQU6rx 3UI["^_$b]NCu%E۲]4.f1kY ?SaN{U'&fȼӡFږW]lK;=Ï`°$0GquY,9rk4Q,?4&b= .g?S#ֳo'|t0B#ӑ)M `"^<6l&R y){RA y )ʲ>?oө5e)(#LOeFY# ubsg&Q R CR$. \J1q{cv!OL≬Gm)6mMgd$69?FI:)z~:*v ׭ 'ƻI4{H1Y_uqb`x4*׸0UiOIÇcz.'{?C ;M̥~%|Uq"NM1S^#j aPNрޔ#Ww#0GkɽFhV.KH:mqd3q{N8sZzt;?]LpE/]77}1p!o.7)048 ;=UN_4%zִa *͡M}Zhb8&L͍WɅl0V23ҽnbܷv!@jl?fMgu5I&!WtV,/a 76 z{g~~egq͵Num5˹&{O12>?lN~UF-B:pr4r~Z|y/~ө24` Am P{W !0&E `lCSho&vzE .`ٷ/[|o_Qf^>2J^'w; U/~޽UC}Ӫ;TZmwW6굄yWؘb? n9CI Fi^$ IşģTw(he r'HdFpNjJ@_5ꃴW piS{^y>6VWݢ ]uƆf9Z "ӥ3IDJ{Z4pY^1].Q(bjM]8I'fqRV`\897ZbCRy,B*T̝Iޮ\b_|]Fɔ坊X^nS<Dnŭ5I¨0j\z8XނctR|^Ǵۃt&#5D p0aP[ٰfl_8>b݃+nzYp"ڸgЏ_z.rTl.h}eXҖ!Y7M.H=ut*l/yH|d9gFpo ]f[ FC_"g֦c)WV`啧wm$G =v~T,r&18{8GXS"%0ݯzfHQFD[c6i҂HJ;Ou>194Sƀ E\XJ&M"Pe==(w}̅cJFEQhTR<`PjQ>HJ9ÐE7<6[{z\h˳[PtNP:;_M0qӢV;pj;(^fePNg}- #{2p&Fe_!$HP.qzؼ4{ya'qQ~M6Ə8rgYe ~b\\UWUN(Nv9f.6l?G Tkbz TV5-ۭ5z>i_)Q:OnR'˧EF$瓪QGׯ}l.^+N2sW546cl6^ԃvVag{8r"E5+H!)pژSQ?hmj;Sd,h|L&Zh$=#0e#"[:2=DUXҘP%KG"]| f>BLu3avZGi,L4)rO*K )M€%Jg4@e 1q-a"ͷȭ@.ؙT&HgSG'¿LIExb=6P5 pZ:; ;\2'O3z0TrHֵ{1E- 7쑫ˉ/zD ˞%_y8RkQA{|ye ^"wq?9X~e7"BZ0$K"_DIYO9P΍7h99@hw$^HY]H!-H.j ^dJx1TЄMk}HnHdr%7$nL?]>9.ޜ.!zFt2jy5?~h~Ǧqg'sZrg 8rl]ZCuf.=^\Fq<:&x`Mx\+Kb@*s!cp(j-=4]!1vMN$GV本FQs ͪTcE=j#-BrRRN/&=ݍE},Q=C24rk]bͅr$(KeEPb[ϭ˦rN lL=rwf!,岺޵_|xsYJbZMOQ64UYkvp9_zb]7f}G4ފ4_66nMySzR ޶g)TaO3yڌiϲIy9؜z[|89fm"\YZJH "[9δ TFq\nۉs+|h4$@31@I-בrkϹ`SF>C췴fjG#E"TG#u` `Q8O` y껴f_|=:53*kxo`5d[qޠ.SYSnz\/ÀW4d>d=%K-5XUS6JBDPo}o;E$3 H6$j<'"4(O88XN`ro/9Bcl)/RӥW:rB cE0*'%rQ}5͈2} z ebg5܀j`\*o(+{eFDHfH)Ԅi@QG)YPxU2ȁs{归o>;D-G5rO XѰcI*Hj b!%rrYojB02KYN:F"'lgWB?,q% TPexKH H4{"ր 5XOV[NyW$Iƒ*@K" 00LRV>EE`' vfzH(8WxHz,e@@JcxpLSF,!ޛAp"8vTo4Ԏx5M%t*+3(nf- %*%-9YR#DsVz =wF!r;Ktp㩷k ؤ3|k #N C3 z6~\[vK&()|)’B*(͹NєI\8D5;jZ{5"jHpFD@# Q(c F@ICj|4qGkAs<S Ą!J<(ÁDF#ՁHOtaz g˪sģ>Sܸ\pZ2kG DhCst M(  lfj[Yҹ73gXÔh'5`M.:N.\ eEU!6uN[lW$ۮ6`v4{TL/kr8-'mxIXݞe<ۈ,Ni%-JP gȝ "tJJ`O]7o\_9n/MqEv=㋊8ley:!P׿\>Mߵ޼L/'y%ꤸ[`8oR&[C,[tmv'O)=w}hӵu 3?/=^IƆT:P \{ZPOǶzœS{s]  KJ(9eVC.M^hdF7q?.{ { | L=@LjpI(o"%$aJ `VJTD]xvVHŲh~;zs78ya*c E7 JD C5 $8H8g$K׃Dxiyn3% @ 4A;;*' gF)#E,8F pA7H%~;sҋ ?$wʲ[lΉ=r}vLĂ8cWȤ*xW@BM4S ΃@O?g:_AtA|s42s(K> IxH!D 4(sCl[omHL'7[5q̒Ĥu&H'VF" oCzd|/ q3p}o\XYC24{-IOm8J(W< *'EU|k[&H<}M\W?yw/A)_j⽊0E/MNp?cq] Uˤ|/>EcLPOŒ;=!!Io'|9M?LrIӿxTϷ5rgGm/ _ˣ3;{?4TfG D#>!0'ÊK7;; ;\c 8:zwʫU\qJRNoPbp]&<ς).J竵ߴ6_2nG-ǃ{ڒK+/'_z0?Ÿ@(v85N -H;؄5t=Gy?k$,Q+nrR_,iv@Hs=?ݜ5p5 S>,~_cU+شYx${>cocaUt<׾];Z x!?]ǿI?hݶ# ͬn.cVA3lc޶,ԪLtUwc- .M;z|y}8чC] ^P*'v9ud$hBf4^Jp)3Mp8tprf[YpS{!b!" UJ OJ449gM"\gj{ʍ_-keHȇ4MvWYJr_ $J#SeĶ$W3s9>B:AX`NYTZ>WWQ+Ȋpc.kˍ{]7ihM9^)7oqv1xѥ/XՐOBD.D] EiTibk˱Nz[=ض?jߕ+ '^{_^uW3@A۵mX=ro-v>4=luw58Z^vZl*pQjIUet685xu$ssq ~")Mffov\ hyˆQؼ0*Y$zr"|UC(Pf79͒ RNr 'tZe'RIuš'6p˪ZD7dL>Iڑe iee2u#!ntaQB;vb~ӣ5kEWJVT|]QτWQ{[ZFR!fe3S! )Yc\s*λ@o仸y`q5=k_S+}%/w0 >=lx ުN^֢4=\W,tP)ZUwtF ѾX޾ynk{uMAf'=k8MoL1({*哢u0|ҸG~s,xeg77=a<> /՞9M ۿ=^alO1faM)ClnzdiķBӈoM#i74AJe$hF|;4F|ӈoM#16)4F| M#iķ4F|ӈoM#E4F|ӈoM#i74Fh7ئ4cflӌmЌ}-{w;v&y omM 6A&y omM 6A&y 眀Viy&y omM 6A&ynMmM 6A& omy omM 6A&s/y~NmM 6A&y omoOn#7O>) Sx+iJSO:OP%#hg:/BpneCFTRu "0Vig yFkpDzbqg4cLʈ,{$"灁hM;'%୍a1@ȨUNbΣM镒ˀKGe=zAe/\c"$VidYHKIr_]M\vqlБ\xGtsGA8ͺߢ5*3/EauwjH1w0 }S ^?uᨭ&U"kPR ¯do,Ŷ!WBZJ*^i-ɔҠ @7kl4Hbg. T]R)EEY.Uw{{uppԚAWTuwz>+6ܐKt0 `u'e;/:`ڋ||RWT-Vݡ5Z!vbv2͓d`C˝kII%<ɍL (#i*qV0Q2#XGNsM$a*8t6} 0'ǍB-9UUjy܃Wo',wqh첲#}pT>+|fIޖKиa/w>8>M;mG"ΝLFL3)A%P*s|wFsO(20阂 2K**ȁh\a2Ev4)94:2ȉF\$.2њ%eŠYE &TV !{±wAQgn9Y_T1uw'ӏ*2̈́HD"<ĥN=XZ< ېJ^]LNxr#W0>Wk:;j:CKӤh7ss@jɾ+f˾"lMwƹ͏=NY^yԫMejT~E^Lc'_ e sQ0齐MM5D#l$l%CV2 k׹՚ۏGܙNõcӨ+GmݢvCY^r'G!5@sٶbD;0!DBA 63njOqq-krլdӸh+Elq3ZIF \uƣ1r(#R _ޙ! 9*[yE\ܙ|g-ؕxa-N(^Ng_#6>A^?U?R4Fvd5kS\&իb:˓M|ۑM'6E*)ab7/}s>,ٷ˟n |O'%o0tݥmAxҫh4k& Z+NJUJ5-^=>M']N VLƪy~̙t'{+ !Sx1Ͼ?q;Lغ[;#F41};XbYǣFO<=;=tq(׎*ã./rݨQkK{ZHXc< /doz04\Em4wi(L˝>\RױiY .e@h<99D[4+Kd_h ?|<(|ww>qG(v>Y )Hx~z@?c>Cxˡ%Uzqß0f2׷:> àekI٭3rrG{I$C N֖qtz$cL9D ;8.ȊOz2tI?n-tFJB;ٶm)_ó CR]TQ "᫶Si_gf;S^&6*bG"&8]MwwvjSym ϝ.<+`U+ye nt:-p㕊yxҘ{;;5v1 i/t~b<;0z;^Qv`{)1b9,$.9(QB=X7+мE팕zH+gyf3Q*V(X?LмH2Qp c>ۃ>t #۩&y0;z|f䂪QԽt֡{+4u_.b.&N]+͇򤛝]t>7+_\o~^޼SWM=L[݋IfOuvf@t>` 6h!.foL"OSru ߅ j fo7?7~e(}.ةBI]XQ\\[ * +hVw߸p]s'a_]\>ja^{(JE(W~k[bIW͏ 5wݿ0X~Ē''05ם : 4WkSE7mG+9إ|.XV}xeUAz'ѿMl삺)}_G>[R\z6OiSKчdY/} ʉֹEy)$mU9 ] K yi^ P,f{R*RL@Ge'Ѱ XniaIR 1E\DZBLuCYn{U>X53_YCVlFaY*el EedlJܭ=e%3mzc;ct{t'gXEꯜS&7A5na5Z51~o? ݜ [\$oaHp.v ]w5=w;nSy;x]^8zV_L''yjGGqJ4Ѱb;3A4੗Tbŧ> ž(=c 'hƞĞ':N%ht  @1kr :D5۶ֲh Ixhk`1:H2JKb{USr次[txu .>{ūOǦPM7DjO b|ơ<^6iU;p>ԣЊe89@e^/Yz8gLsOēiȞG.]?7|{c浑t21h=xK<3-r؟oĝ^m/<-oyNo^}Ԯ0;y lig?42xB GǨ, ]nOr]n"ZNDWͦpҢ'bY)cNx#r$UbRJCId:Ք]J⃋$|)291I80AP6j9Gtz-*h~ԀYs W0~7L `nK3]xU=^aYsܨJ{wl_[kszc3;Co\˫!i,!cԂ\+t-/I$PZv9(Tj{FJ ڠ)8R|$F.cD[[>xj9#ʪCTmr*IYk'jqitSyyj^vju0I F~ѥ43?F-l0^Gff۟r1i$x#rұbjVO ŪE@=d*[Q=/Y T΂(AY% tX󅬀^(P,$%WrëWuFˇrX+}6%)VREBAާ$c0I8G;XչCf;aKyNztc <3@'˄1;%4O6}\/_s-lAL.AR `4R=fP+{ ccʚv5eMʚv5- Fo3{-b$2i&@ IO%(`2k(E] ૨lB0E7j). 7|cl9{0g:吲ƽ8Z m˱;j(ȸ-/5R E61|S2@$F]@8ްĦ8C =hh&^@e⭷̋=(2}ѸicV tPh>Rhh_=8c t;oY엘R;Rz 2i?V>x@+ ɫ2XRN5VJFה+BR濁|= ./ia1=^,naeqlzBK܇\#Z͛k2z~iS|G*j6d1("8J(l5.l8䦚8[r.1 R-ms}gW [; Rj ,fl5c;L6Յ.Q>."-{k}Edßf-3E,-8b1aĀJS!R)b(ڞWNP^ TMel2 f >3EJ8_=7%QɺynsƎF1Ek7ڦ6r]T2(!(#Yk'L#; 1eU%_˾>,ؐ!32,:)k2 f.saNT c͖s>HQU폇jDX#Q#q׎GTJ眴uw7~ `FM6j(J+m[z9)jƤQI_ Sly#뙘cr H1ThU,nj9kO_ud8[=$_g]PE7Q/h2dORu6hr*EPk0(oB/8}@o)FevIgܫ_TS]Ipcw~}R41k,!#s ZWc=]J D!$%kyg >F"J3aI(·h5jCHbE1Qպm4v:" BUt>a .& gUی{&Zf3KqM~}&|-f WltoZyL|eSB@ŊUS 5 ] bo1˓,7M޵q$28H~`f":OgrdaU $4 ivtUUQ 0J c9x(ց rDYKEIvOvAQE(ٴ{fì,TP܂3xV1sp E+ d n8`)w*,Mm@d2[nxLLLT<e))E*ngn)!!QJR(*ƠPF5_ERjâPjk8Rj׹'8e&aIuِ,gyR qQ.yQTߔ'3/|a\6oM ;ͥ < _*$ Mj;cqk[T(g\2qxόx<0]=7m5pe|cDfl^CCjM78!xښ-SZEuSN|Z>·d6K8ok%dpq|!7|nJ[qjC_h(&FK+)dtz ڕ椺Lմ9L靟_Ng.8%삘kJwv>[nv@Kys7 0n$֎jHuÈv1f(AEɇMf>1{dYvT6:u{Vk]Ic9 ϏQ/(JE#~ȫ8ӥ la~/.$Nۭ'ϗoҟzHL7sKk &,A]G[-&Ui h8:㬋U!6<ƕ#*.ΑI~wߟ_~:w\ӷOߣ>2u$Hx6 {A^=947ZZbhS&|q tk޸ 8H!J /ԧIܨ#E'N*b$`y3 Dmtť"cLRKJQzK¥n矻htCAd.pAюkKW"-> ޯӑ_g*Aw&JOV{hY;c3ܚYE\(gm]B :hmb ty`<;0I3{="0Qy69霘3r^.8(KI]=Y΁(ZqҸYI9[8YJ @9 uܮ K(( QL &2' (3B"7+!=cZmT}Kz69n&f<&^/M5&՛3ί~sUV~M>hM9}W-x%Zc5Ww/^YOU,wmLGӶuɮ?YR~Pt6O*M)dI|4RbE6e^8sCIy 0eIű?8d_9d'@5BDŽKYΡx6wr']H.rF%|o ,{d13 #V4n ^b!S:kšJTn|p.nEF\IQhp(a-lɘ0ي,ylxl[;z浯,іhk[{^ɻնsRG/ʆ6jZI'W[[":dAU ܗ'Z%?ܺؑȵ?s \QeC)Q , < /{`SFtkuiyY=(eTu6  _[6e)K8~lYFzP:˪=XvuБl(YSK~1($"I"2ZJjN#FJ9OGFAFf-`BeGB,ŸMΫ˦ݰfzr!-u>u '9e-댜-lmٰPpWn%,s8c1RuX)w$GYKRLPo!_HdѰLO,i' 8ZrjSVgP+N{=i'dEJ?*:L1-=oQ,{PFke(pVRA"ȹӊ[],2*aݱqxq+ y*w莹CF1mfEvMI=k%dnE&lbŬR&/fN"U.'V t~  %mFRݓx/i-WHǦz Ow-+-=/zWjM{QoPţy mN-T789{tlS>Z'C!{[;{ABVm)Ww|c m{3a8'lZσaSj~\"CEVF{mEӗ]o! \ / ' ]U #UVOGnz(qQ%7%fA Ь`鈳zxKUZ9 AN==--NC(w93dzsw[O<}])TpŸ&6M{~u鑭ݷiz/n6+&s8ս޽Eu&f<+.5"kˮνYԚ]STXHc[f[غ}~j]O[o?QUg[ز4zmx$%yEK-7pͮgfͣ@#x+֧_}F7+8{ ^Ҝ?U޶yIk\ ֶt ۲foi{%KN-m嬇"f׍}<;}ax^k#X *[:)lH\g{1)&7k]wǣvfbYnQi|4L9A8ͽyKQJ%J".qkU9]6;]xR%B&Df,'-K&"3T.W=QA .oyX&GSr{iz ]T09W«>y]\&׸Y%5@Dy=YQV@8ekOBֶdW Pb? d9 kwڝLCp&CY8ϕg,9, *2t64Q260 3LI S+3ٻ6r${$G3.۹rO[kEHra-Zc=ZtĎ]d~E"Y"j2m-'N )roŢWNs7:y?OW"ps%̀M+ITR\g*)@(:dIMJL1qb:eVPiTXNCxA9ts9Y.}5rW%ꪛZYv`S슠k=v8>EӅj8b2*m"(7uHEt gyj5O'3s_] gùҎg L9cY&NSQ,+3h\ VhiRBRIq`R1 hv71҈x qoQfx{9 z)shZ]`NҠelߣ,P t!n?,]NZi`vջUW7 / OGϖn:A;ٳ_(J_FkR RQJgyot:gDq`ѐc#P۵_7s.ϊ絔3Ece}ꕂǫ4j.|BL^i+G]_lb7]G kM%n_ B uP0ν 5oIOp0ߪtWzVޯ/Eze-|.* zVz*6GV<3nѕu ,Ժo2bpFg}ut)iqa89hD %J & "2"[͌4"Q1ҡs]gY9SLᩅ~ ˋSǛ/2J+~։jy̵uFQs, ѡ8^訤9f6E8E(\qg: gciΤsϕE3eA/!hRJۈnϴ=ӑ-}碹4oa)!pEeey\RWt)*[~RV1 LfA,Z(}Ri'ؙFY.RVR."THLґ4ɰ 3 8WQH̓)\gkEuHE/׶1A+-k}y,)Km`N%B\x ϤpQ<dBs\3#dD֭WamO{Ĕ_hYBшЛj"2_lwmq{qKLhR^`k&e"J*zmfKr|tmZ/0s0\&@ۅ0I2 ՁFqqoFя?-9ܨ$emOB%&=(5Vw9 '/B_ATRAAga>h\Ȟj\{T%~`?~60HEw|4(NtD+B|כ^CuX>ZHf}+7է*н}8wTy {5*]\ͧ3seE/|sv eRݻ{i}gҭ;%<فI$ޢ<'7_بY$A3}Q1aB:9t\>e}`S_ gߗΪu11!qVȎqvE;Y[^gҡȗFedg}lRHQX31͛/°뿯˅+^}l|<'녝d0F7nI7\nz2r8-F50cug7k$ngIG T:;i_;9;1tڅ(D(&t[]wQ為:Wƅ̈NNQ$z1)C1D;)ΜcD_Jt>,l5g6[?<Ωzq~|[>ժ'N\ e`S]p)kDNv%r%J"qeeUzd˽u}ܚ6a3ͮ*YIvw 3bpƑ['/;k!HCr`ZJ-I,S>U#H2ԂR/%I `R1 &-X'8FP JL;+(]m49 ׊y幍2.4,g< r!T!iwSuL]% $i򗞮с&b Je%Y6ړoaJMo(D؂֊"=59@e(SA!-ݪBRpP**bJz!#Ccr@B e "t,:\"(, 蠱Y3j;pĠT$UMɷ&:ʩllK@dj<A[**dY){h*3jƀ(Q.(ʨqƂgH^_Q@x[.HPSy[gW&[`R«QQ@IBR m[J`_(@H*+9JOQq@X*+T햡 /\5$Vc+:c"& LgwAZ޺^Ko3fUy' c@JRv'A<{0- |~:tj*n>?swJo]5q`NL$ >:b‘J{q:2YWҥ`ruA"X L͇ɠ<Π;l _jTcA#zPr (1de+ JPaZ*172A)ЭLQx nEe06CƂQ,TG''"4+JXm))w2<}XwuŝCLm]LՐZE}m6F82=nbTnBl \*@QR"d2ƲOZ%tMކFL3"C(jC&@k.g@!t9hs^u@XK4iQip|Ӂ$&e#s4pqإ6FuFM39(k$`?Aj؁0"ep% ~X0FaE^Y` E*C DYl9i M6C:jϢ;K4YF%Gj3Ko^5 RYU[fy/z+렱^Jw afs؀$>&J X[ҹdJac-Asdrmevo\ӘIm}`ڢ+Z`8FL-FTah XeVn I<5vjOs7zS^jx4G _6D,T@s1%%C~A5(QR(.*PP<* GTi'O H@vX 7nkdXllFVO'j'$Z"bʅs <@bB|;?$ ^1 `\ة %p #EiFF|Rz i,:27hâHYq2csJ6HaV 58֨Ur@vCO/t$ƀ?!],Y#9sP*m}Eh o:k/5s%z1gHacX:dС 5!WօX~)Q#2`pGcqV;ͨ>rI"WŪR盉0K1:c@MpIDa8ubɱ:.hI56X~VW$<Neg,2BTvիmaV{q|;O@*-kKy/p\{?/SUޔܵ᯼kS1!dg{q LWgY6zR";><ﺧGQQ\_y?솜7>Ŗ{f7԰JJ ^W/^zr%V4?yyaޚpP~ ;6W!h C]Kn4whm my3;_w4]-/omj?b8(:YC#U&)_Jp9pBxX18{T&p[pC( cd^DڑNzeJTJ&O1L{:K^ƽ¼|Snd<{՟c ̢3|d h['=2,V W @aZ4~ܠЅn2RB.΃wEgx眶zT逾{8#{B^kwEOeu5f8 {wmŴ#Om6jlMbdJ-Ik3Q\4&x$C|+ܓ=-hkT v~f7K^Q ެad j4^tq MRr[~(@SU]:*dtJq̋4:+W-#wmC 5@>bfCMS<`^dc%V| (r#H62e B*Hؽ4Q,sЛ_4J4_*hO?ڮRAt4If9`2 Sͭ*E $7 k`(x<X=2@Veq1}_Aחm a~65xqy^6-}dM\_|w7>!rp61whz&Ĝ7A9_s;njT RLXxM; n؛:j&?$HsƇ{5T- K{jgV2J,Vې_nLb1ަ:"͜1 rюgC5vNp1{;DžqT"ͻvw?K//ŻK凂72!l/1 6Τy0+\e:1-cY(xG z{N뎟&شg'-wL P2i yp %~uZJd9}tDtCtS>vun~}& ew%IݔnYyg bV`*K7Zzשּׁ894zcK{2O C6pVfk%jB.}w?4{7 ̩65ks{Q7n!ZnvUfݍlX߹\U@xnvk(C=nous1R |7kJV+`Y__Ѭ:r6>Χdxw^k.RTfOuFtAR'/S;2:J@ہ5FEURv𡣌$RYJ \bUY r\!P#i1|fԔjvD>k f:9R=; ʘ_ɧC|~70r8Em~ˬăQUEmjep6Xe;@rjh N*&_P'Lg0Pq!M;ܦRF.ɤ"+6#KHcιɪfpY;iIqi8Λߟ~Sz&)h/q|⣾ /)ό+gX0g:\ a_x,OQzѢJIQzZ }z(M0r$4FoGBwe;O"$&Ev1*]av/1. 2X9f4_uD?F;8w?/g^_mE%ϓyK?hvcbr:]]Wi~q=V'I9Uw'iOJMĝ'e<$[Ь]j}FWy䱍 eQ\cÛ4I^=I7i/޳q$+FGK&= !m %);UσhLlڦa[̨U]y|;+6p m|?b]㷻$&>[4ӫ?!t+f2xqvtG F}5=]Y τ݈1Ki&z M桽0Qow6C|0u,Gf_fs*h49uṣ}YL{+dJ4M ңg|z`@*+ aװ280 ).2qoisK7zFH*•1t,2tYOPWÍW>^ ^6j!xKj|GՇT @up-]qZ@ fyevւi0;A_iBvp P{ Ky r:&|IRM>E2kjTATTӱI쐓:j]cS4FhYn,Y )_v AcI60)QTRFLYqje iP rhcPKy  ;4jD$f.jNUJRIPIu*F1ήvP )K$mB][#F ѠȲH72C.5@Pľ BB^2j"ï"I`YF4HV%DQhr,heD1.@)E/LNmw1f$;.n0B)<#f&n5+N=/1NYpCn=G݌?iQݣ8*ȔArIܣ5C -Et@PBQ(J!?:$NBdC;!TEr9c1psrJ1^X 3兢0/#/|)R۬9li}yN料wtslC T#xBn,*d|>Y ўB*F%EW6M س,M.$L@pMb`"|L+]:ϭ89‚y(\v˵ea-\nBA;NsT͸ S^S-k` YA&LY~T;-(!= yM-ȴ8" օ6G:*[:&\ 1b1 G,rDU##G) ZQ#_;KaǡC?|m[Ϸ[EsNl(ڬIpOf,_R'LFqrkHzdAe6Q|`j'4 .%PN K DL$@Re}~jwé5e)8@(,qTxѐPqƝ.wti,E"B32FrmO8Nj"0aIA5jo1pvRsӃHDz[(!E |[iQvmoYN wQ$sm5|ȝҚM`\i4) B1Ƹ˺ GCD,VE1NxiΫ[I ԡOr\Wi RV*h"SڇKpJ B!^QlFbC ]?`x|AD 1@dPxRIINIw#%v xh@=TQO)/hPB{ HS{Z @t, ,!Xt+4sB.E=ڿBbh/CFPj"bg\u+?~= dju10^Wʞ81Iv}L {Vyi|. / 96-N2y~f]VǤiǞFoȿR=gV9rЅ lsfX_W?bI/ otsxA9PBZA=n#`-]8>d1 x$I(F6 AKKD((4ؠ>b}绪rӫܢ,W'rWhoӪ~5Mox2WdmٳVa\M1T4œƕwˆ'ߞ,u{:I:y_D_x8w U]绸w |uMkV߫.o㴂{m2OhNycNQjߵfch!*\JOOTGQ}95a~e?o-k7?o&wzD |#Jy]5 Aօnr"#T ňѯ7`$eBHȨX?Ѡ%G +M';Y {񱃓PIcW"b9¦pq  Q7W8c=rV2E Zٓϟ p3Q ?E3fK@(`< YdR|-oٜg6L=X ,j%sژs&v:Iݒkυg󸐹lRs4P-2䢼z# Pry3CrRᜡ!FIFl670 m!njiJa+tȓDHPy9LZ'D\XJ&M"PebʽP3s+1%rM4AEJ*)L I0a5Ĩp$\\a Ƣf;$[z˦-ю<:s}@Ir Ю1r "Xk@vQIETK_bܹ/g'HE@)4z!-Ž&P%@CR$!7B^H\kJ=!&jiuA3&H|ٻ6$ $#!Cl/;o9߇McTHJn{Cq$$-Tw=WU^s%*tiَcU]gf5Ixy~kKkJĦ7^tJw˪=N-:ztQdwED~ }0aT~4DIQI)%3ZF$HkV-2ɵ)Ti9b -]gtY5畂4}Ѯ~쏦c,ʚW}F=4@5vh!S*gO1)볍+F.͘w0#Obil兑ux׿فM'q6qw3ZnLZ@F{;h89tk*_s[߶vp=l[BH ƑVzeuypyLzq2&"* .$dܓZZJXZ\4UR bϧGn bhAl<%)=WU;wpD lhXZFY w*h%{gLuH ÌМMgXoq]~UIi\˹ۛw*}}IBښ\<=b]jT>zv"iocTMe:iT*-eɹFJ 0ηW' sB=?yGMgxf9 *UB<I(jSZjfN()K. LD+T4hiB$ L n9T[َ( =8HD/8NL Y΢;d55ӤSABJ#YXizNic(h[%pWAs~RptL>@(TbeηV %W93)˺#*`# xPK/YGQAܳh+Pwtӥ6tް:6P5[9c :8D[]iQ`3x.|h>aG%,{èWo/EKQe 1SfIB#S^",g O%EsL(QmGs2RȰ, Ƹ#aREb?D h$ #َB Kφp2UywrړM$݂=AgGG 1_bDͯtYO)z;j^"iTfm+tأRA`ċ6SK(䱴4/0/1F$2H"'RJ9q2LvZj̰ V#t' A!e!xR$g/IDkөl5ͭH@%)4Dcz=R.xO7QFQ^z^(G\h;`W>K ]?^vz)⧛"h LL: ]NÁLP2tu& Ut/]>+AH,UjaY11wc.й6g3? a&\{̓`2 S!9\1[ @6} f. dcQvufH2љPp$aPҨOFeWiU-$ܓ<8 h᳣*Kd57YozT9{3M: L.p RF^9 4 jq>b'\QxD_:g0Mij3Slx8~3%HQLsШ0)sk(7D W*1,Rg2DrTh CRA(rb.Q tFclW)qT= e~dRs㳵z7^e".vcl1s&xs" ={6/km^3,CFsu`>̦anRvCDE%y>c@a1u-!gF7p o?~(9f`[|5o`Tp~n,&ٻb)<û7 Ye٧:[9e2e$"_jY"M{112 &?Nc`\bҥV9+ڋ˂іp<3B<8ic TțR`#HPj Rs@ef$j5]U=)VYޜ g\Ytt4P)0. a!𑼔K~L`D-3㗓K[ہ^Wq_9cُ`(]eaOck7_edr9_l_>T!]2\ ӳ, I$VsEO,3=< &ѱx;+y4|qt-iBrg2k0E><=y6SMhBY>bno'"6 hwu>^7̉8!\:`o^0="AUo nT<'NvIߧKf^y ƒSA1ߕeiiHTCtcuG0ONˏxf#w.pVlwi,Obl. !?7C\UMGEu`ziF^;6'5oI3(R©QGI-k?0%Z#w.d]I}+ ~Xʷ7d4Úc~f*TCK@mT.Z%dHn"T:6гn1XԓNͶfS=5T K66=ʰ* Ĭƀ,R` `jA 1 ^HC ݥW=з36DˊOhbv)X^Grf.V݀Een<1(p-iaD~gbN2H޸l LW{<5tJ,dp-C^nÁ8Fm eshu)X▲NEf\Zj #^kK~n)5 U-5~MintG:XiȀZt^AlAbZ[ (g;0N-Ě| j yk̙lH  /A t: `DR5 Ý./"\xIHPcj (kvk<o 1 XT& :E`FyK2ru9H[92jI; Ƨhr߸56]Bl6Fi\`TTOЈd Ǡ߁]Xjנms  ѨTl9@wAKbi0t ((_L(=,=K rr,ցhWۊR;Ң =!ZT(5b`K:&m)4;V`VQL "<)k@+cq9Xtԙ  38Z. Pu N7o:4u aTRCX4:5ZVNܭkl-ћO0MGfEVh pN=pc [[fjCcMz`ꣂɷŪR *_ِjO֎Z^`5940)߽7߽@itŨ{HbU 1O-U0 rle/FVυipV ^PHj/=l)05gK/Y`Un;]2C.V#Md?E×H*0ĖeIJjAH.gܻ9U~AOW,2"搂RH_ u~BWt- h"?[fB5%gBop9P_1jWbҳrlϠ+tV.}%h=Xw >$-7b沴/ox3,awBN[o7}nsiwu1:+M?ef^c5rʽUTWc*9 WQhJ^1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%5\!ͫ2\epU Jٽxx7\ WZp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J Wbs4$}M+ j Wph/p(6pG4\,+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp+n_ z W7Wc>/ WNb#N WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\j#iC4?.xE/_@= {Q j@O$~zs_ڝSOi߻h gWxxDsWo.ͪx˝_2gku~W=IV&%[ZY)tVgP{0`nݷkW~s0ydDVi?ǁpkJp߹5>%H+3+-jg@zX鸙@\2dʀKwGŬ ʛrJWoA]+񄋛X4#>߮/9bޗe>ClَXN΍z51q2ET:-*OTMe3&'2if[pUK]; lPE~syu>/{{+|85U~ξT_㟿Lz_`м >Um:$I-yG2&v =l fj0Liᜮ[ngmC?;q$ [onS|ZNE3jM+^ЃMjQEm/Nǐ*ZtG6L394,OB? `b]D|m7̌iv:mϮvퟚVA6׿s֝*ʯPyȲA<?<()(F4;2iLEDiF% \ܬ:eSq1O$(xtFM/ 5лdˊ%#Ob;geK4<|{cc?ȓ1k^]F[xz~B6M_no>؜v]hxo]m*O)4# <ߒ>1[z!KY<->mbd8xEUO!'7d˥F;ҹDuZU3^$3N 瑫ŬHD#9k+f}sXTloܧ Oon{G,7h:w|l(yPG6f]vH(cU,j •<[L6#{ܧ6??`p)w-mI  /=Yu~JP;IQ$EK#i@UUhx|AD 1@dPxh]'9%Q?jiFJAˉ׸ *)8 Jho!WF,HLKFjbXt4sBNE=Bb\!"dްȄٴ9Vb!;qϢG 7Bmb`9t^9z}22QB9aK9ϩ);ǾtNrv%xjl(D鸁\-؜pgrp~WQL2}hC -P7ťlpf_5$yUk _sS+ݣ%WG?W\䨒JU+iqfy6៕wՅ7l 1F0='o?u=疛.Hp ~fA>8sa+ dHl aQr0ˋ=2ba%p,|84=^w'N2}!7Yb3? G>w=JE-q~mKXK[O_^_ccnKe &,k;lPh8W:DY+RmyT+#vw//I~om(gowQv{DUg M?EGpg追evG5Cs ; -[}ل06yø?kT~DF 0 W6{>OJn۪#F'^*j$g-2DU!JAR!+i[MRD臫'=.e=F=?n xZȤ!"*-7sjr (5.N6 Ӓ_g'˽ vokM<(鶏ALzesly ܩ1gs,XuyfN;w+oxN0RGtθҥR!xϵBN&[~_e4m{P8ݡP?͗|_/I$[oan9" q/˂σxTr~X~+>eJ)Qw/Y O,w]Lu۾ď:u8k۹}j~Hw7EoNj8܍^PW؊A<BQ&, rL".͵{TѺ)[5:d7xݩz@_W5 w 6{ p)=GAy>gL^Q~an-sc>o19wg463|߽E sZT c]R-=FnY뜓qB"bbsrvrSeN0_:$PZEvN%a[&26dnkYxF 'R[3m g:I͒[g9lRs4P-2䤼z# P }2oSᜡ!FIFl67rͺGcĈc% ް5Xv>+vd}v194Sƀ ".t,Z%kAU&ΈbIj_#s1%(rM4A܊TR<`PjQ>HJ9ÐEw=QckpY}ykܩ޷s=wwO d˷6dIQ$țMxrc忪HVOfWJpR5)fjQglh0W W碶)N#oq ۮd'.aԋ(R`!RhH4Yj$ĂF(֮Q%Rg-ךROmZC(wItOZ#U!2k2J[LԲ:.H0ùV0-{6T.?foIFẟZKJ]YvV޻9exδ|2t|SK~#($ k#j#R&sJF"F d (ꀌFSVIb\\w'tPmXpR|XwnnmԚQ,s}nխ(պy(O^dEoP\_A1vyǼ!*vSI ޻jWwzi~٭Aw`#µNnM/4-ĥsn-%`A[K-uw= mH?̼Қ݅ C9]0GJgW&Y (ךY +]+\ƒ`V ?m1 ׽/ GmJ \!UM͡)pc"|F)\Fh->ғO׉F}ؾh| -ĉFTs$ u/EDN^I#QpprR@rULya)/RӥͻGrRGΣqQ;CaLR ]Rq *Wi;-r@fO1o7_izn]E#Bݾt]WmrMB`-YpT $R8\~NeQ6R,z+Jy8>>1rT#Ew,ϑ,pkEi Ri,I+ %FٙT%`騔!\zϒ@t1*9g#_2 I[*^ޗR)jí&(c5Ļ$Dc'h ٠pah?;>ѧyW$Iƒ*@K" 00LRV>ͥ=P5'tГcĹ #ꅳVyc.ui ^%{sJښJnxmS<*KaL"pۖ*%%\$@.z*-"H[N/W%0_kK'ZfyEK%eBi5w'˨(geMzHA5'^$w`cJ?hm<;d!E#t7Vo{9IY5)gMȑ(M@!䅂Mlҷ,8y.5rg4to;l+öɅ)I[7tRsfA`>ģFߓƣOm}ȰGňrK@ ʰģn$i0\$Q@T %pq bBQ( &eF2}GIOAjk옪iE~uc[*g9fwJlT.5{"Q q\6pEq(]/tie*]a0>^H6mUWcs݂nIXҘ\ ґpf!D>ZϦ="}{W,d/EK˒M@\'aQe&a%RRNˤ 2\DUȅsa=76p9iI@e$gSG'"[˜.3_ {L;'.X]q/1)ф1^>rz%т'ٸK #Rjm8nPyHT3(=}xiExIa$SI$SJ "K(e(P!NzT>%v\ [pd%Q)69k5|ex9-ڌC1Y i dJx1TĆ g>Ѵt8v݌y)nHR|ri~Z܂𲋛NF*{|Al죋3Uμ9UPԞ^lgZrg HQ%DK#ZX'Κ[BCKy!ۃ!mWku=m&msIJDrE{Su-4/Z0,-bXYg*#iLnGW0)$*0P 9m +FRu9ɍϹ`SF>׳nfe&PV;)፧:c\k\y+`$S&١U-ȏ+O [z@ǥ޵q$eOmÀp^lpNn ~J(!)RőRCy LwWUU|I@zBndž:LՏ|xb:Lx u?ڃ.73 9+/`4# YB"t@PnΎdU,E@V"j|R̅&0FAshcƄZ\r[dm{_{`{-Lo8Wǹ/SMܧ$/ #EupV0p2@2WE~@ Q!8Rg2Dj!"?)Š]%6Ʀ/Pf 1P]y"I:0e_jxi(ex )xӓݯEqR-L`6RP|Z9O^ *!Dua`"sΕb<~,۹+TҦB%Ts܋+T0(^*G(Y(g-"0bp<)DE kn!\؄ A2 ǨU1kQR1>  {t5Ц/DIKkmy"s+9u& BDͬ%jK2pDX ܨpvZc 3AThªq(T;mTM76mp>2tJX#PրQrnh傃2Bk. ' R !<{n9дL Bi!e,+ aF* ^hO&\k59y<D՟fVc ҶY_jg(Btg36[KOs!J n`RbrAs% @nw@T;E7} vyၳ&6s5ZnPR:(02n4H68&|{$"\9p8ջ^m0t~P+7֞ >\-a!9bV %\y&F)".)pJJ o#[6ɮ=LPkr#/3Vz6kQl&_o;yR` Ht4s7(n.\qXt_S<`s-:2ё_Rjw/hvlLOev3,lY2֯#ߏfk_Zޠ]͕kzLrB($<3@cLWL#,8ZI}-P۶_Ts.rEVϲRS7_hy ^} âR2&P@_e_&~Ԭ#}IWѧaML5}]AZ[ b $<]Z5z_Gݵ>1&!1>!sߢtW4I/Xz)^ڶel3A=ǺYЈMDN@\ΘH"!.ksJ L {)E,THA99bHztDH:#9K.1XHpawsjXt1yj.}<8~G(-I_$1/rCmXjKcģ. A_H/e10J JU/Wl)N1GcY `,"r<+ 0H|r*D-)3kH:h4Xk"(y(v !'Ku'I} H a-jdvSpH`Im [l>ɆRJ< p5# Wio(Y"W2 z;PB"B#G`a.^octڱ%&{%LY(sMB7 (_xr_eq_s4PG ~:.o+7Uo%'bdz㫋^(XI\@0&a2e(iJEf%3TKKI ~uq=~M[fzESkꏛ[f/{"hw_ ^ͬ4 6aآ W/}uW.{i"hH7,-ny2n2oTf045_L_Ug ;fŜ7NvŔvYAbh#Cm0cxëwbJw oҹ<8"aǺlo6IL6wibcB\qL;|os7٘)Or)-M9W7ۃ+3xҿz?x)U]r煢j6h^f^< ̈́zbӒMe/?bki6|I4"2FdGhp)hǔ ?wE *dJN J"B)]de@e5tWx]Ò䙓1JPp9D v9:7TF"*-ݩ`3J/q4nm)_oT%W.SqauUҝ;œtq.I7o"B;M ot xy.F|$w8TOŠKଈRry$/F<_Zl#\mr\xo'wJ5^+k)L@yM>jb w/ fV4sUN۬$5fv3NK^yCVLLrDZ/W_A}]^L5<*;7?|^ڑsrn+6u(ڜ-rm|!~帣lݥ ʛދʔʇ:+ cg-P5PZr;Af >!PTVR J%/@ ʇD&#Imt#n1/)ڡړW,Gwmu`aU^`!:&tgRپ7bBLx.~BuhYyhzu+yg1X ^JD!2OgZd#| ӢSV[bT۬a#mdQҝ؜ dފ؎JvT>`DG)yDc~4:;=l``sz*@du8G@o%Sj'Զ^tav-LJ[QY(Q!m .^罟f2@:nJNn0Fx݃uqљWRVwI¼TيOAusۑS.#:lu+ U0ǔ񱬀- +r|hmYmwΊ!砹|}pcߥF|g„OmzS*3 {M#O?kv.{RL0UJI ZXK6:Ytq(q,nr]U>( ~r_ȷ I=ߌTCW=[*2M^lw"( Ւ3„j{?%Jvoߞer=?3~P^Rnp3ɫr$N nmt"~g :|.N~{7oaLMa7?Ws[`;Mf{Mq?56}5I&/>kkb?Ѭj) syZQ|ތ?yݗI.`3JG kJ2^|e|\ `V)?1QtfR-F#G!1e^\K6++k ?|sf`RTI|7$覿UTջL\ۂH ݯ?a,pFb,* 3\Lރk$ZkDd}^@uY1)1_ӥ1 UIe*)9xk~YgӽE\ o_Y$\ɗ^L|qIG|Q7ť%RN1z6rh8{B}\RXJy@PFQFJSFlP;YJ) 5W)ٚZ{`-vg($YW5 @w]_Kjy9UT{lWQn&xtr~9˾y&?ݦZ-kbOFÅ&wRӜarcAK-4oo;LT~x(ǣ^*>O46(쯉v`d!h%_Qҕh"|!K~m[_RNʬOQW )H*/" ˗{4[ثa8ހ- i)9:uvha7J!4;:`^ : IhY#N#i\R$c!/& + Q&%0Aiun#=SY-dꗗ/(_^1f4z(#BR(zTHC^09]Ќp$ n#Bazp+|h.z,͖j㪥!p3֔+!ל -$p2Fyf:+,}F*ھTަx:2JIJ~ % _@ $UE3B׍g!=֍> :b %/` X%sS~_8/+X'g~x[mEsjQLFeľ6˭fBH 颒te>9/wijLoǪ;kÁ2++rCmHY8+7 CZ2E 0Jrk[PBB o8\#A{.;xIjdPsA!L9o?BqDȾq;T‰S5}kdk5$T0jzYGܼĺ Ѵ?wԫ,|6Զx@(k 'l9`x(&;p ] $=A|Q|ܕa̐;ח=2-d12 e u}9 k}u>-8d5fŞm֗0:j%ї سCz g´pv$L\Io0+|P"aV5DqcV j$@ %Qoxc'p1qc2\Viuu&snkYMNFf>MO/fOprɩT͗Q,,lɒbCY+C[ >S:N6f7~>.$TSX!p OM<+Bˬ$(ڛWk^iLOD4HY2P+ QimF"w_q,VtrcQ T}8ݪQ$)a3R¢_D2]h '_?ͯgj1x?~ Ue0r\_2'0W-AfHI(5\i?=W(u4+vAS{NO;TP[ڙO*"{M7 -PPi! Sjih7ϮjWgk<8Kjg,W<̟Snn#O ! @ IôEGZZn%^'m(*3lu'# +5KPi&VR*%$NN;|~ϓեkEO]&h>v8OϯKgg.4ms+q^48A ݿh+ }BWwui Up ;e c/?H~>Fo~a&c}<񭞏 F1]e%h2^,GScr.gp6n}XBݭ%] HXc~6ZXUElGGh M,IEġmqڶΆ{(9(/ <˙0d@9z3br~uXкfc^US\9UN®t1+cV-D"?ޯ r/ʽp+rWP43D\s 1'c4%p?i3Jֹs*bzOϕퟃ?(bdU486}rDY[VGZcYɏ/./?:e{ uԅRH]4T:Ep%a$bڍfR+4F"*XHeEYG7sH ١ln,Z*/fb#C0kJq9o+ ZY)4%Įouckq꓏!\peVHs}O  &@j* I FC]\"I2/N !.A "nʒ3$I@,`z= ]qkOB*!k"'.r xέJ K샔?Pڪ-/+80qf+@/XxwSijS@0[c.E89sDyBj:`-U=l%gN> K]UϞL" =tlo~1o] 1?{SƋ`)^f|&C0(昦 $}qGkƟ:7gM0&Z*U,Jd gp~,C4Eu#ٮlUrW(x pВ1{f8/8dc?6 佱h̫alk!~|#X%QFvBKUfuY'D: ^o\%:+Z()Xȵñ>` `ҁE\QŗQL ق˖fٜbRgp:HO|!vf2 ZNF1jhvCFjdO>>F (T Onlau k##+ =e%ss 1لO$Sp0h@i\Mb̦eYaD]o[kkyZ17K5aBȓSa}j-4{ؒނ]`P$iH _d<ֽ|LTWA Bd.4oeV(/M)t }YOQZyNrJawYj0#z/$&'\Wi'"GbkKӇ:ED2,CM56#AGh%D!#dbL=0w1LJ\$җ>GQ|Fidr9&kJ?ƵGPm'D[^NE>KE i/\wo2n1`dKr.I%ro-UWԳ_gkkZHRfdl"ʨ+AK 8a4KvKQ(Aba]ȹEH%廖}]LiqK&klY'N$!lsvdmQ(QWk4AHb4ѕxG4pǯ!'ZJ&Qai%w싯 gn %t tұ[>u)iݵQE"Xf Lb-3iH)PH5{3ES>oyTg5XK̭Z7)ziwPmY 7@mP% Y/b *Rv<S;TNXW5T­fP `&3+ʴBhK5AcšylBfʜ֦&͕d9pNrt!4GfX؀)#$ٞwkw*}^S]wJ#;mL_X}PT@9:3bDͶCQśkO$S̝86rYX0stgam"L}DH @/R{o_蕿)Eՙ=,i&` kkn\7582UMe7M)F$_,[DJT9G4dy^M1-2>3b04وamQ4 p1mmtƍ(D`٘&FZtUwNT]3% H1tĥR8FTOC+xA@oZk$t0zh %s[SF*W[1qy1]#٨16qNso(B^YUwbAYU{1j+- })!(}z׵"ekPֿ\U,KE(KH;~tj$ Q}UD7\&$+G?ɈB? HHDkAI+IXB'1f&V|Op'9IT6ILEF`)"v:KU Hg4*MfQ\5ʢe=Da,W~?88U41a@\8Qڄ4fVpg $]e׏ְ*$OҥIE4*#YAf_($FNPjd#am;PXrI`\zgu ?/AH7wWQp@?Կ)9nRQb@: XyMAIF`eF;OjXCwC1H2d].:J9`ÿ34eh"`8Oʤ.5#kEՙ5C.u1^|:g4㹻 QC~: ㇔Q;0 P2-׹#-PGnUaX\qȥ/myEi<[d\%}ֲէuZ>l Tb/L+ dL z^u%fmQcج@ [㿵՟+klFILgw:_nb%ar/&wTvRȣei|)|~{,{P}B,.D2Yjs_]Rd鷏9V\EPCuU2Ƿ h G@CߨęV~ɸ:g`X (fJM8f/3o}@c}V9'mlHԐEgxycahߜsE,cg1A\dZC{;:#A5P@ i$.UwV0D)y{Q8Ri[S"6hBT~ҺHͮ#zէ7Uæ,)IiX"[QV;se*"mwyH$xNKl{\޸J2զvb1 : KueTa2Izyb)uY $Jb`˹ݼ?&k\^Pb}yXB5:Pe1iL@ăcHڪ 4՘!(P*}vHj%p&fy2+TT`.; UP6e{쳲Y\}~O>ʚVY1W9x(6J sg:*Q4Y3v>/&r4L7+`KX`y- QmTQX]=J n̕5$ϒ,/dZ3.9*e@j%%Ff[NUbP'_9PY~Xk|SR?fT@z:6Im5W֕τJ /c_5CE.ezz,Y$a<o& I0]sEX@_SM&xө G}{~ \L#)꡴[d8ŲB<|U` )V[G5lY IzpVaNĐYctڵU 7d=<6}|j$ԡuY,C4!s 8O6Lbn' qlgV'Ÿ䬗F-GΠB}l-V1> Ee8\c; @XQOq:ʵWS3Rf H1O _tr]eNʞ6+j_7B,F/AI'fOwfuۄ\Ok C;U#k9Y֓SzLAkk茎.ALJ0 u.}Lt|Vo۴v!ȪG^cQ?7++^7ܽH9k؄0}@7۳Zk97ݬ5>ݿ]~P}c2@c)f .ėͨmjr# 6i9HC&NdP2&Hb M5L69{ZtFm5hqmN=8z3 rʶ̩g6k;a Ie 3izH$/bLd,q?!1~Q7 d5ࡈ`Rdns.rss$FN.gdt MȈ|KBkvPi"0ƳGYO T"#sn dh02Sf "76&Iy̨F^8}B|@ i/I>ÇaQY.fdr&1ڰ?xqJ/z](FAηRh|D=3 վWOyU 6-2CI"MHB!ޞjGC.jgF<ϣ8iCGjհ0l`bf_wb2_ˆW|;M. pt`Y_KԅC[$ b IPԟ/gs҈"̔3bq _ϗE*-&sOE')ˇel9-2N)sg*ЖX0'ih1_bElqJWEƅ{[(9ɵ$=7IS9T9AI Aߛ ^6htҌp6IzAPxTf hTz88*p#/OEwF)ANiXuLY1;,TZd B=Nڦsi35QAx&#\ +MqׯU 4"! 'fPJT$ib ᡖ؏04NLQ-OS_vPa4 }2 0R6]8f-8ya.īhV's EfVʿVU?d0#|3j-uxXSk)0Hhߴ e :>ȫ7 g]WD{̋K9NV=h|x 2I *ED) uNHrڻV jϬGj)Ge_Խ'!"#L==CedW`f@o1ج s* rm=&HY#x#{qơkϸ0SS.P]&yNfWFm%*$R<6}|8}RNOZd\g,32&oK?._>w=FAk%r5IxU*9NCEV,K&$3'.JJ WFJBCB6NKt7Tc (<+"r&Y~=]u#t[eʐ8 R8hhi1*+xxZJ*C1x=gY΁%KSf9^9#(富2ksLaݎyCJjUq°chDtk] ҨRd Fm;7cqs)Sfr%}R`2/VBk[j^wMU75IF$va GV\߃M Dû 7z]cspصp5xֹI]jw-[-¹}x+NV7a~NCF*,D܃A%{ꖼ5~ \L4\=V1Bmv)}ݥ}@DyV2ާ~#En0:@GB=Q9<՗h8/}8_XfW|a40RkAVBr'ziqd7ļ:Bka,k-#HݚRi 8 ܊wIb`n7PP穥[m9+ F.6 <Y!y)U鎅@L쭡#?U[3 q~l2&q#i 0"©_BƊ>.ZQ\R-&)+JRRʷ$\`C{WWHE2N' @D r~:G 4 _Axgvّa,vKd:A  讖/ $K-[b͖-{^ƒ VU;"kt"x2o(ը om z5g#TP r^W 5o$<٠X8#$Ǽ`/b^rb(F2^ /~>ޟ᫩֬=K02yΆ^H1M;Ίz@ IDxmSEwۨnOL*rU.D~/P+Oq/4%zBfO^K[dlRRc2ʞS{_:sXSqZ߱_F > @ # c@˷Ut?YXˮlQ +Nmt91"뎹6pï6AmM(n?2dE\ MFΤ`:T{ `c=3apLOcFc9"b 8YBT1m.kXy SJ Jd~-hj)'8 ޞȣd{E2< a$N]{=59QF,7[3T8#d@{*"{`֛Ϙ͗>ì[N GH,y>ytG=?mJ73Zv~01ӗ3EURqyԇZ౅*i>1fy2`pYx(9zA]|$h.|6m’YX Hh8+Mu8W5#O !Mm:qdf,|zjp;Y.5ѫU;}rSI%*d/JЮd"PcX?zs΅leM Kwh0,48O㰁tv7|˭:8L~dkfD/"P;WW5DD79ya|'if%R%B?.+fs3d=7 . t`*/en˅UPd\b/PVpїݕ CT8o3LDLIvYRw˜\(w&w*ذĔt6pڈ TPִ $Y3W]'. -IiU}R| ؒUX3I HAu~EG^"ߞ .ՑL0 )N4uB}DPa.p>{$\mljCeSV 1W7X1CL^BQ\Bq`*l"ŒML'<62g[Ujづ}/=UCڣ,y=Ըn3Sl'B#,Xkr3Ɇ;aZ`MTˈݣh.S/Nj` pĊNp`b=TiY w0qԱOk% o?g_t:~ZXWV/R%(eQ3 ʴF8y'&aB50^x{L}l[bHe1.*9=b%4ФBqWJ5 z/Eb_J2}ZǞLOzW PVn71=0C.É +Fs1m $ ιᔣv}gNp@16)7$9F{J͔u * ZtzZa.T=QDC!Ŭ2ۀB̆ù;-RJ,AI)sI&ICׄĶ%DfqY=<{ܡPNG퓼k@950>ԍ"YjF_Nex{\(t@H-ogtSJ4 $ƙԹ Z)L5G:͆n5"J. oӘIDJ1qH "G)Lƶ'Iil,%aW&{ J+;}=ە||n \xPԡLQB_504a2"{ϋ=1(#8`^aKH0>ab'dZb4Oޔ S!vXhƔOBLE)Lpz"N0)Ʊ/\5_M$Ҝhb?~T59ABI4 R*^02Š?-iÿ8~{5Syxǡl6C.k` (C":9jS*)NB'p/ R'6},9\_./R{U?mri&S#{TsU.jb.eiktLt .-9x"O4z.$k\U韅E5-ewT/F שqeJuŒI#SpT( Ԭ@)Fe^M#8'0 $+Ltk~N(}ч-v|NA?* lJ*(qvu%!SX+FHkd')e9+@ _xٚ-v gN (}w?vY#F i54(JQMpX0[} F > GXׁ '!="o:OT۩W< i"TdiM hx _0hha?2zT]yY}6XuYL֟Eaok!lkԅB50VgK|ziIl *ٵԏ0tt>bm$K"/AuޤX`;tY,ڊeIKz0})Y%Q*$S h+Vu]e5bgAXcdJ&6nޙ:2*BOfAڣoZo݇jc Nqp`bGlk8XpΠL?Ć4\w( xy!}2O<K=m:WA㝹 > mO쒧C6JOcCC7rN*K@!: +Y]H%a NK;=Z!tnEB; 51'ew$'CK28HsSXj\Bv+ 'CUu!^͸$Z^.VEgh}552a)0S\{@9Yw$mXކ52Vrs$Xă`5[i]>C#fܕx=4dBɄAg<tF7T Cf3"/3VpJSaҕf"8,gg* 1M1S%1\Imb5GP!cRo>K47Y*m}ް+ҽv^|[Ydh3~Ή"%*# ZNQEi.Ѳ:^0)J%4A-"FNUQXnиSzkpŇ`7cfgx\.z(JY#/3Z^[3af9XÀGq gZ $sy!Ff"w#Y92Cbobݫ]d)|2>MƏaAT9ssrWfMgwg9bSz6M(\#P0_a[e^e2ٹZ/ᮗ`0L1JP.(ԡX 8/rtJD mbeވF=E3bkkj[Amw7D* /E*zYd9a^!Vc3 ?z, Ysl&_`59a9OLWqCKrPw`{% /|AٙRwݛOb4nQy !3+hڨ~1rOWl\#%` 31@<'`j+8-Xm,i)G6`4e+#̓\ϒyxuR3cuOdVLc4 _c>))4ĜNZ>FOkKv^Dۭ 4HbٞKM: ۫BUrpxR8myU^&iYwZkԔt_5hWqQMٱog57zuq.qG,=$k.!1ca%onuptu>~u }(7%k'lʷ4LE/Gw?R.֜÷=~X:xQ@`||:fs8:a>t_I5?[)X}k)Ufte>qy om#7K fwfrVnZ~ư3`k;[R} j>r`j_8E´ͨizyɫ4/Snu*'c-jVh^u:Ͽ!Hs~'x#7$׹^8(0@? ŪWZX_ RD2F.K[]-d_jX]t.?W*h>P 8Iy}m&A٪)u:$ J F2di#~_6aŢACy9dW=8J>8Ye!G똏ל{>>^S6[FQ0^Cw+˝83Qb#pBb0C{qccY/01\oY$,61Yd{PÿCXj:+PfaM_/B^aTyg{cvp_܋-<Î }`Lt,ɝrGH <8n:;S1wŃtjo]ց)/ 90ΚFl:]nmƅ,񜐜%X?>՘hF/Zn̑LИl]L#85p|"ldz_BMs>1z>B#@J#L'oj*h = ]:N&&mvzш$`|ģ7oOrkLlpoV %7+m1]!DrИQez54~7οL&ײַϮ{}r|S?g.+ sdڕ40 Q[K H"h * 0h8GeXz{P`J-?V,&5KYm4Ii3-N\7"D46Kn-\!v.2͘rԥQ3VAڍv )l:tVPv3ה"U[SA566.-咪 f^oPUNŀ7@Iȋ܆n;4 2Tu^J[(|Ų}yƴ]424z=kAڞ 6trF<!CRB}7_1mN>سo"l҂>yh_fnhֶGʹ!e'YkȬhXA0Vxևyb '/C~鲜+V[K>Q9I&ha\ z!='-54n:1Id۸bH5{#+l;6tkΕ탺({hh9k6ba%el͝-b{sD4 'A!ƊG!Ea|7J K)*Tq ?,8ŕ_c:SyUG)& (j|3p, +g2m0Tif0 pb@(pAiED%2CtG6_ >|$b@BpoJS5 A?p@[V,ׯ>`7&#-c7xI4@Q["&@\Hl]^ҨeKH L|_nC>2c-"c+Lĭ͝}.;V}FH]|MJOʏ+>dq1~i8s|o@[һd `➂mۼɗ@We45Duڴ3<7* #;u8ȒH[02!S1w Ux9C/ɀo$45'jSe$|66\y>/ǟTC ,T۠7_Zm.Acos P ش8cF42Ęm*>M`eODBZE7;zPt쮪4f! k9*Q);K, zyvnT(1Z"ZhRnOL_$j?FX*u7qoCyw$qm#qGbUШ \s+!Q-bMP= ; O2Ei87Mb%KMY;HK%UcqlDhİxJp-3$ZkE !Dz{jlJͧ)u|;[,9o#*a Ht(obXœaʕ:VX"*314 ! `hJhJ͒`%w|?OriBHq"bkø䨂:Ph6'n%S6d$A+tP>OW vGX9 q *$Z"*"Ph(l@ 2FJأؠ?qRy+-}j0T0vo1ժ*I͹5*fqxo ڷݡD@&ˮ< j2tL:/gx}$ƛ{qG_ _)H}ubWߙgWN_/W8}xe~JNkX{Ξ<_B!΃\F]B&y(4\Ck3w?c]4zʇ]&ucZѷz+uz[ơuWKc"c0gI6i:Lwz} 㱂={AVW*w%"tnϚ8=wmHt $Ebf0{_xı$W%KI˔2i $6Ū.Ve{ԍC.$9vpnr⵫Lתejަ.4cw=x?c-+Am1Y?zHh<|ݛo6ئ*dc̩*Ό=w-z8Ot. cNXm aQHEAـg%j-}R>狣|e'Zp|zfz2 1W~f_|4+9IjE8LoA%2!d({`dsXS-2BaFhlzC@] hǧspQļkV3F΂il+R)k'SɓJj I"VҔ(2=%XKT6֎ށw0Pj@)r`3OEFJf$FQ9>-V$'Q!Y^R5TuiGɧYxg@ej1^gQe;x v~YU:EA'dSkRTX2K~Ol2)c,wЎ΁skV8ø,,3S0HI:<;_X|lcQE*i]-:!ji+](r.l-%[GkG@];iy) ;i#Jn,a@+HDѨ?//еV+)W_ˣ Fܶ=x)cQI5+f*O}Lp9>EV2[<!2Vd"zrT⒧}[!ў<_ed2AhӔhhIDVqduCM:b,**6~Lb NX%J?,V:A@T˨:gg;6I7ټV?RM9SWLWtj~x=Ɩ1m[ ۉ1=$%2رhcbǢŎ0vLF؅=E[G>cAk1m%Ok+[oU2 HZ(tdo"*jQST2mұ,eЬP)r*R))^$lUc42Nz'-&۝KF'YV͖5'(YHe4K0:v#2w,4̯OvE,f,cd{Gʡ9J^DY:[K (+|JoMtl59~9an WieOvͿ.6;x|ϲNExs`sP|gCw!vȊ(j JxoOd|7{JOg狼_dp[׋pzFH22 x{h~CbCf54NS2FfħB̼"as`"ls*JQ_M (>aP{=v{1l}գG_MpQK̓[{(`["*az"yˁ&COATpjOңyN<77朞WْzNLwo7y4`8WMyh({|qnЃ)𡒒 =>3tdcd==H.ց~zZe28:if%RP:L)E3U z&YAJF k(PgZMv(ݎ*BTB-'@_R,o;PWY eFy*aVGFYbQeq~R9>lQS 8w=nEfVo]m/ )8y>MWvkuDz^μ+_bx5h*>8=ĩ!h0"3ѰUHbY*BF+񥉆M4l}cEzlq9?!B~fj2(FU˙Msu֢~xp2<9##O{祐{Ud:=Fc E hgSpZ҆9-Spe l7}Tr}wVuQW(&8alJ2Xl)cVڣ KJ"i)v:J̓ |E}j'ѲH֡ODygQȪ&L*AHȬDd6.Bj HcUmXca I^I[mLXgL2)4=4Y3NOնvL `4{MZ%hYz~VH˙WG,58]!EԢV,eG_>Rt)Uds*k"TJ EXyj(#hd7y="0I+˂Vq|Z1g~z30"0V#Q/joUvx.̚1 !Gk1,`rT H. 1F)SS怵<]LFddd-VzD%cP["cv# FT5!$~㶼 вaK5%4րBK8Vǐx0x异RY|m` [$"lzSS rJOUBY*`lJ)4>77҃-w}{S iSj{[3Ki/}O|oqEy;\b(`ktX_+V7I gɛpb_iߕK"^_s>G o٭ܕl~]NJH󕯎~E$@~8(wm3X2VgDKvpkֵE{e럢Rvc3nx]n$GÏH;. i`f1c`B\v/n=QJ%9ԣV0tIee29$ta;9i2kEMY9O^ko `PID@ɊkjL6 Od߂5l( ,njV-,-SQt OnvjԬ ̳M/1R0.+{ )"Ht%R[bAŘPoK >9c֙eUVWH_“S&UJcUR,/fR;mչ]R#3AZ֠|/Ўd#4eŕU|U.%L VzfYz= qUjM^RT[OYMZsMSI3<jv,Eڱh<*N+ ծg^܀˦vhr*\(@5쟣r]uj^jq‡j6֎H;~~*ٛՇSO6v1av1a氋y1e!r1#Fww»-6&) TE<6xX,PSiHlP2R6*]qvZD%BSˢ,9%g?:lSy[^jW7Q;Aj`'#CBJ (+}&4Uib6AmMY%B?V|^Ӡz/&+O[2lvЦ^qS P;o~5zںXAHJ7VkaI#a(s b!ivrˈ{] 2`Kj*P`1`NR&5`04SqсZ"g)q^}}SDZ3g0>`:{#? .-> XD_޾ؒx=T#uixY^d@efzSGC3~_ 1&Xr? Ɩ-TG>wXM3) x1l!:Q}Wcj~D3O'J4֣P%(&JVXE#3!ƾ$GA,~玶 yEx=rXT&lVZh\=f[)T& S,K&VX3<*lDjE!F7ޜLo2#-1iAF#$ t WyY^a;Op `Jn`7 ,~Ϛ>yTAZH^:ئ1G+0+I^v {i5:*+zPo0*${]Gޞj?~,gvlG3iQ! ڧO)N5#b˃pwJqz]Ϸ}HMd1ILsMjw]%PN;3Ӕ$I(_U8Bcpkc%y2"H;lQ\ٷPjq:Pg5qհ_%|XoFݡAN?re0S@1W@fc3jvS|f!i$80IGk2.q.E V٣Y'RH\oz3b[(mPZ7+ZGͮ%;@c9;M|>$T9`>B3{jo [9 @Dי ZKj [R#bt!lǮaJbB{ޛ:i?gHߺ4?"M? =h/+/Ru_u"&g "q/ɇ'fJL1a  uLjmM9Fr - Hv;'ݨ} |*|_[7)Joxt+g9n>|o^moS%y_:ܧܷw^-( b #9MN>e' O?ut\ AeU$'tzB>1}8TF&g1o+Kz Hz§mTwl5 .A6zC u}]*/wywkq_^ŕv̓ho1LEaZ~wu?O:[_ҽk.-~z#^eo llޝlLʟF13mY:_-͛Kׯqxg^Qc(@c//z#}=/%O&U|?O`ωYw/u8Ϗ̦vtZ@4vUK+21ƄU3JP3^6j|יlmKؚEF|-:'Id"ɐ]tA$uAQqÉ$HotwǠi!Ơ7y&1hXࡿjv>ovbв! \csf|zxe}zvqo`:3]@>^w?bמosNhrE,c3Ƈ4"J^!|xuͳcj NjjPZ;)V:'}2:띭5U+WjYYe[S׾} j{>VKV@ԑb$JIF{.Lُ8 uK?hQn/l\Z]EŠ+e+&o&M$^\sqHZ\\&>T].͂dP Jwfv,EڱH)+DXZ*c.3]Ls=U(oдG@GJx<=Af4XO ; @wFwi pcy݁evRrv<jݎ0 Jk{+ֵ$7da K5ҩSϬᙍFAf]@eqG!+Y?e7}3"px}^_|`Fa1NL"ý{bXR6ސ>=d;IxDo4> .y6b]E.gV+9޽%tz$ңZ};kV#CgcH8y/HPגqΑ:\3SDj@կF{߱d=Ejw7;N>}ڈ뒈4 g> 8`뱫L$a%[ )sX0H')2>R_Ev:b8%@OSGď8y_bV>nM*=$ëR1 cy@[$'l{ݧU}.{bXLhol.=\t?TO[UvOSfB2;'t) i=Q,@io{@3b̆8tڥ!%¨x-.jzԨ}O j(l@`^ |oywNoI ˩;Tk[Ngњwe rdC.ىnJa[Ō`E=cȡ,c=4Y&3BBMm9&!YJ1 F}39fXG k4ŴQ?f/vڰˎyyJ.c9Z\y6dgP%.u^4z+Ʌ9gq-LZb ͦc_oߍI &9$=Y+8q\^ӯym{=PYrZ#\Qh{$ߐU(tcHQF4VYD & Y5u#(yQQV S~}׋re= ad`ci;ӾûQ̵V}Wj,F&5BCu J a ͨ< b:D[C\3t/qPwNdO)I^ezХfP%7*SZY#~jRtjWW+#%ZF3WHBaZ淜؋22&8F Bh02 =Gш}Nb~L>$ldMlkd cd;':N25r͕7?wrRA`һ$7UJٙלsnJQp ٬^F)Ňso5t#m9FիWÓ71^2>R`U+et"` Z( u_u>&p8Դf̳ FM$KuG׿œHiZ+~ (?I]qah*i{㧺#fi7H_w}խ#d ѐ+^A]~r`کz ^J`2dLg7*3$3v悱OEQGV^I.I0o2 'Tn $K)z;Xs4Un7 3}\Zl,QG֮fmmuZR j)HP %xӏrT/K q G/><%Fho{}/ 1y}n?I׋›f='vwW(_6<փA2f˧6#3晸ZEEぺ#Ʌ9#}ٰ9ش޵f9oja SƬωM^beӉz~ |jss'̍+c)O/M:] >~/NE˶LYXUw7 :µ:fS0+kBm ؆Pm Oƞ^'Cz0Q R4̍ qi |#leͥGopߥ:W&U'&#])v7`Ѝ s[稝lplsX#%֙yC9I՗F0'D`LGtaԟG.R ]/A__24?"^Ns}LSs]K3 0Ϝu&|/܅ 6Ooa~|x_"ߙ-of>ة_-ߦaW~41|5Db쿤JnqA?5RrCAnî<-"0,cص$ΉB^(fq!׃܁@ݕR~@Tn7dprO3a*ۣqk56_:x Ȝٟ i@;^sT2(P#h朶~N[_o)חAk^HJZ^X?睽F.<a[u.6M{fk:?Z|Oh1SE1K{mt̉6} 7^SqK $',$_n[$M _]uYyxFK^}ɴNeSrQlPl=Hm@U(f7&KI4~N4iv ka sKE+^.AZ >*r륚$-3{(B}mI)p4=Ik_#'iң$޿ b0 #CJ$pcPc]8N7>r AjEw1sURj%[& etsrWdK)>FJ}Zwyt-BJO{$tMkrN{dcnL;u57#G!n,= j1|ʰN=6‘iv7cM>&DYGWҜJ6\30Q.vl /px}lv߾ŜYvѕ$jvp.VXh;t7 q{k2}=/qԚ艮M|\oWя}5qyr?.ዧ`gƸgA^ z]obaS'#ݾ?A^JNZg!jSy=t#&@KWzV& zFCr\2Ab i=o o E7tK(ۡ]tJ"!FOn- $S~8Oc.i1zq͌Ʉrl< FFNT8URPPI9 J"-aQ̢컟w= .~1Dpdߙ4P2;G*;5a4n75ZAO(l Smݒ`\ksˁ7gxd7ރC7y*C1Ildl,9tux[I.iJO nȯǀn7ĹmtCMxvwzXk {p1z# 9E{\=F2cy&I)$u=d`4ا7ѿ " CF'TW7)/n}5 WiW 7_(҇k_C~|_W掣iSZq3j;ոպM\;V}ul,+ui$iN_˨e6[VGg:2{TZ- Dh`k!( 1`7FCG!x%mS Y%֐|uѯ7um1Ƴhkq|VO8J=r pj( ͏ΦL$Czw;~s\~[scʽni^Fx*#B\s.~K$Nk4v7zz2h|OJ~!Shs#Ʌ9 $ i{Q+]m!*q3l$2:e b؍t+Ʌ9?{#ב@oAr亱H 0I?,-n3AztOO{.eȖ<]>"4!Q'şfRu~1IqP {B<1=)ޓk;tf&V.0u;٘WZ+8ͬ`{z'ɉ9[%K3S^؀zR5$Vj UBri͘MB}Hfb=iz"w\Dk(S r8;Q.2ζaK[iVaDh?y_T~A7nMLh4Z{hEWCp^P;Y?ة%Yege35^>ԮSI|ѱN&#D7vw$9[N:[ }HSB0wjwHe v6÷D<1Ru2E ^ychg1!:vCwSD>Mʎwjn ӭ9ZO.v_9 =:[o=qj'/ xZOaYQNAzgdCg}swEvQݣGr)y(8|{d5Q& U?Խ#yKRGQw,Y3Gd9QqT>bd$ԮSi|KiR;:77Vih\vڭ>]ܨU'ɻ:@$L: c'V檮ei˲ʷ,?9ʷTJ^}䨭g}E|_wQd%$ `i LJvrQvj;VuFfN=˦  Bj#p7Vɵz. 9l\ lRؠeYQ'q**9+Jd)!PE,(JFTe>dT cz[I֍s}=G`U{:AbsDVϥSS81bJF QE JT[86PbkͲ"ڊl)o!ڊl,mS)ؘ`U:ű0QLF|Xq ++Dc}iE˨\X*cCՉFXayFO5m"c XF .hˬdMh &䕐~UU,e>Ql2MW0xk]|KqA^Ma锡,(y,m"[zW(-crH1(UGnu:V]< 6Om\b] b ߽:o6le0T~{ Ty+ $H2\f( @Fעʄ&fk|! N- 1/'A؋/N?-HQ P 65./BM8k!JtI&{dS`CAu()yX"<8Of'JU)Zܪ~B2RߪzP$rf6%L̒"5N|vlo:^iY>W~C0 ъ!J},~[tl0*RUU"r q bdPM. fmd1YIQ$$|bF +D&6Ph ֱrC!ڍM^KPB;se)^,OL4[j]*7ˣ11('XcP`8eyO7uʊ+ fY90{--` w'uA0:4Aء*NYf9M$T՞\V5Vu`$؀(Բ *هYTJXTq1χC)2qqkq+9E=̀?/D$9!/qc%G1opo!4,b+!ld\YDݼόX#H:B(SMT,Ƙ5d"9[Wȇ(ThV3d ͊<m%+UWE9:cU+;G(ϮZz?fx⑝xTq?-9Nd1yp}$>p@ȃe0'MfZJUC XAt6Vȱ>%vL$(,sى6 ɊWp ."ZC|iUrX/= XTEh fl@PvG#>f eVNX8;Or]?e8|j%bMwMR6(l'% 5ۘ[y{LJ6~L,2%~~]sXZF4ʗ(33ٝ)r|~Saɿ|:񏃿 uQ2@I~ xӽq>ۤO\ :&5xسMkh\ENы]F2!|7d|{`y!=艒%ЎYkܶ[9NLi#B5GS(>:tО;x͡1Ё;MQ+ٍ$':KyeI =kړfz2VĪym{م)|CFL$䦍hB2iÄ5F4¶NsyQDRVE_3hY+uf4H{3quVszaBC6DzWqVYkgձUA4yΜZu+ɉ9C#~VR9'~Jq|ӮV;g:|CVa@q+ɉ99zwRfw/)kA&/;}{:v PF^6*(nk^ŷc M`p$' n w <<؝-Jyplq /y^xqO?f^枛X{ަ m)M*.eю,g)Gґptع(~"#))|c04:O?f~Ç|~NQi:rT>?l^kO芶{tH>*>vC>~ioegG ye'r=ZA8dip1VUmr7Lx·|Hm+Ϟ_wJ|Untr4sH/rSq-Uq4$HR)@ 3aRfj%T6ԞCOX|:Y?}K v 2s}I|~wf6tHR=QYF4/)kRڳWl|]>kEf>`Aۢ~_~Eܽ_ΒӻK|:-Dar!WOǯ~Mw1w7qa s2l^Kd?9ꙝΧeq̽>-YݚnΧ;^}i2l=OW}G*37:vuq-"["z1Y'q/=WYjޠ9W%8r#ZzmF%cbL08vqP+j:fz _㽖՜ő~[b^G'aצA!Dʃ% SzFb^+Xe0J4(LL[Z֣0O< #(%YV4u6SУ0/4 z 堞_[9q]I/R;ng;nߘ]N"no4$yԮSЧ^`{_~?U7K_Whn>t'j+^&`SCgX\Jb\zoZZFMf|s&Kb6E3AassJ !1beP_ڦ}_6ZDJِ~wY2.}u)&= K5n]7 bS-npla:{YE ?3Ngo~!s^S"F9@TcĘ!/1ZEK.KihՌqduo9&gބA0nX} =8JyJB!Z&vURV`fk \iЌ:lUIw{[fˊ/v﻾eoIt*8:va8T/>Tyj>iW] ׃֙wmr=/"}{E:ɗ૭Ė]NgH٢D"ERE+Q$眙g:mN휴*i$Zhl !D-M6M\elP2w2N'fEJ~#'!(I$HU1/:`Բ@+Șp` =%5q EJX~HPKȥJg=i*s禹Hm"ibVI ^ޝ-4ӯ༥Ƶ%1`uh+aaVRf!Pem 6\t Xaj%A973RR;P"q (vm\U,R;vfEjP(?$jaĄRE0J@jn%Y)A]_CvK0'(xe W-R;v&zځDL2̣F";@j*B;""ڱCsvK0'A*T+œcvHԮI RQ:xNˆ9Wb&U<R;ir-6-֔*ae|_)CtHEeT`V!!EBZo!RAPnĴ4'\Q;f"!} Ӏ`d恶x&cT+Gh-RwV#%Ҕ%]9q- g;{S@d- H( gz cvwlPm`:=3* hUah}BCu_*C t/0-dI~F$ _-7ɵBjRrH#µu+Y.`#Cj"ס*tpkKyd4M}Sn6_wNH\1M%û`=OGލEF—qHK K@(T1qd@"ׄ*T$7ps]4ow9a_^cG5Jm<ܶP5 4o BIIl{ h k,׭mf`t.GXNSGO?9ݢg{~JywW i{uXY0c(7pT=7)7#-4wƷq'C~pڈ@$Q%obysd00Nrt[}/إ' ݙ:sºm|;:u:sL::]oaq,<QMzRit}^/]B|}@0rՐmiE9.R^ewdtuWY|^ϤrHehL N8WMtE}ݴBua-=NN8{UgYʬ9T(yz͂ Q-MRC4,Yy Uފg5޲-Ӊ,vu!$K=g='f~z+ WjA# d! '__79]6˱ܐr=Y6(!`@Ү`LoOUu[]=rn˛D 5.p; HĻՍJ.޾ς./+Io*vG*"T%[MEBER*niesvi#꒬qDǬנO!ka3Wx0E%IP2_…[W\ vB*:_ն %_j:W:c{ eu 8]{mU/bK6wU./ɮ4,KxE#q,:^֐u,Q,0 FDZ+TNUVjԷ[Iq.Xi[o,inǔ*)p3~|Mu]e+%zݦWcv]ܷډo.ꇰ& HVѩpi,$@t+ԝmО rX ;QF-ȠPZX%•*\,TS&HUY]|<^hC6%dO: 2חog:9S{ƫ -͔uVP@.K#MI.% \ʃ˥p+.' Tݷ`@"IjY-HVaZ+z q .ƥN`'%H.J]vyavx۬ċyZeөWϮLy={N .x?H?NOz?|8]1#`G׳C`ofƗwx]ts_, lw&.;?OƷ0tMc'߻~7G ަ<0rm8~}>>Wd]7Y]"?H,g 'PzHsO^x 8&ϰ\vZ(i1'@ X.v=0($Ƒ"BQm,G8r Jp+.rFpH`#D"i-t#x5 *ek6TS޻gu"#M]l R Zlx=W"MEZoB[{NSACyIAd湓 S>w]3j @ MwM!QyGĸ>JpjxvrSE|SB\@K违~Oo7s䥭Jvדv:|S)>y袗pxEͩ7* ?IdU؂-Gl+P߄AQ7K%֏,* ϕnI;/RӣXOGyƠV )1LLet"$6" kZA몡LO< A!)4KX Pޔ̟f\i <|6jҏҒѕo< f_ kO+| pu-HJwK P|"k)}G csƪecGFa"jcI4|joc%fQNe‰DV2 L>2a'aœ hH":rLɖ,zE-I0FĚZQ&ZZK"\5ڏCp#cXLHP*,#8_`6|83}tsf=]nj:sPKJ%YjNFݛqJㄓ&YNjA-vlf,xy8||H,b)` #ĩ]nϋ?﫭T3vv0w4;1Uo9ՓM3d8L=^&͍?}x4-UV˒; ֺ\K/x_31HAuI9n\3M0g/.>XDd2 _@ ~D@LV 9+7Jt8AKD3Z0 \0i:[&^X (Y<8 Mui‹ wԜQO4x> ZCr%sF.{+l# ЕIjdetPknbnrA4]2,<7r|3 g>v]׎x//CpH[K/x¶WC["D/ƣ%pP P) Bmk ҙh+xy~bx :1pjb^^t0;s4졚cw_1σO#\d7b\N[U}}=e݁Vb̎)Zlm+LNHam!+W~ֵXE*jS&^f'݋gmPQ,fy? "HAý CڪeI弊=CeQ6)zD4-S99sއu꼓{`=r%зIy.Nm_F"AS놕&T1F>ѨTICԮ%Ww}1Q)0I&oaA)%-Kø??JU\ 'ŐFfDn~wN>wUӶaZlU,f{7HK/}{y9g]pE agM:p)58K]yj\4R#vW` *@89QhTR3ʼnk`Ukc[3V} zA'Dx;ب Af`HR. f~ܷ="$L 7l2t w+ʏx"Ln H,$W-DTQ1cu 0 bc/8ɝDD@Ypv['o;|~lvE& gƇnb!oj$PD׺r46ޏI)5p5@b7F< 5蓄9ZQWחyщ?-l*C#/$+D7nR͗xW݋2zb(yq*<>:*<9 yk/աŮ[?V\lj&&„IBI`4J0 $c,!WaH18sYL02U m'yECj!! BFN,Z B .b BjC##$pǗH $+ Hܒv;{*!v^<MxLrLwS:Ne©5FK8ʑ4ӄ8$ ih-"Mc1ID2;E˾=Mv 'N(P N͜l!l=l?ί,FnvJUn mV A 3ꮻ`f<LJCL*C}0]&N/a;ձm6'u>a9>UVR%>L"֔{anoէőѰp`)ŷWd~ HFphDƩ̜DY*{O}]Jc@Z~~ VD= *_e6%\;>Kk$?*X63Q *k|Ƅ vS6e}/yHhHeFϨAR\Fvt6H-ϣMUbvh$+mɴ+ȧjQgK hLm{6=(& o Wer0ܛOƉ!PA(E 8 O10 Lʋ CхTaRsT6i=NʧfeS;9ka HԪV{+(\ k|(ʇJ|׵^YOFdlm t}RR_ BiJwBdA0@ -OH!(2R]ywZU/ۘC Ks;2Q$$$"Qp, a&@*k?-!vrdA*gr$Ew!W!KrF_LuYB.H~ VZOy rXgQZĺZx2quߛ]Ob㥪Eis6H哄VMXXӒ$!R$sJ#(N:_its>L5_llJ.M^1Ebؽp6WfUu )\\R dmOB%pL‰AXMv: W=[4$>U(SUHǐQJ>f;q*J-Y.E3 R%v>L+,ԩ3 9e™2ӎ?HиBZ9\5}|keW2rR_hqðM&}å'}ރqf/}ơ @><$1F`y8O1\U'ϿsV]Vxs<>5`PR j=TWʄ(aN٨0*5BdQDM8 JH44>gDO 'Kݛs`v=:)a݀B 'H,%㡭;X:HQ uJn~0ӛF#MCu@nijSǫ2B 8i#б:Hc!lQ&՟gkQ?͗[uO )\ 9@5(r>P9`k`U_jeIzBJ+ws˩px =8%ǓTyuT#ɵi哋$qH3)XR7DGIM>a>>yg1LvɫfqoO$-z&Mw2Q*TJ'ZXx4k>L; ~%Ew&pfoΗH Ͳ@A0|\6z|`}%K#8Zp-7[6Kd8d05p Th 4 '-_S(&gzz%r<|G;MCx.Bf?sjb3( vg[\쮾}afd|lxCt#xyxv9Y|=7AxlΊ_ܛ!;$ mGK[>ƖQHh69# Ib-8ZCGJaR*RiCecTҁu6ͅKeNe`ԛg˭&dP tJ?^45AF~FqgiPjvǐ6Hvb`e Gxr0A}}թ+qI֝*5V}\qW~4Q3I)-c0bQ"`• xf:s$YlU30̑Kw< V *X+ZEN褉}N"R8-Y@9Ek*4h>`$T1y=P9#Rb947 _BZSC [ /o$B V(8 R ,H8`Qq\y;Kb%3R6J՞qR<9aqu+7]R9z#j~.)(48Tj*܌>yu/^K},fp$sӀc}$0sr& S+ZLMDij>V.䌠0/ =ދ fJ 8SrH&ZЬqAVf?^ w뷝k8&upw\4"=Beϭ:tݎ+|tV(D >a 9okI^,|tyڜO$F@k9ghtm_i񑘍Oq")^#ɬr)`UQ |]  %Mgfڝj9/K@c1V"@ "ZjTQq1'vvVd ^ `xqO';v0]6f"4b"!$i-%k4cП{{pOL90zIdt]3aRN^r&z뵆KLebBe21-z굪S=̫vvAO\z `E`)ź p8vMގT%ځSR!u2O!ݹvR^ν.-5|o{םW<6y@u/vZtV^wł"A ?x$݋p{{d) >]2y7(;faY#_!y QvN g1&ĺ_2u8x!=M@(ބJp2㵠f[5kRs(!9+`|*Vc` *|imSy;ex|hpk5v?:ү`x4 Zcxd$ΐD(MMa~Wd.X L؅>n"\qztu(rx>LYn4*ׯ`8~tDN>PŠ9}be01HkD(ԑ@ 8UUkSrSw){jF,4 r$q.BK͖o/ t7fFHz7|4V۽^8eU.,M醒D#Sx7\0I Af"$B3$2ƅ\ HuWq 06+/E>b А(aɫ?޼|ϋ>IAzi7c̆I%4@Q-I$)fP  ;RP[8*6,(M!l CF1hCh$q#\>%N9OhvOZO\%[4sg)}'E `e%)[_ݟ^v/!Y{ľQu_9x{=E͎ }V>B?lF٫RXy# ۃ@(7fc [%^BYFigz<#1&&lk0Wɖ^- G=ǧiJ4Qe:' *h;NXF:SjO]=ejQb4 O/V\w>K JJ MjH#Lpz. 8+n`n֍``w 9ŐxA(ѪW.>a"XU!6ƸvQamPin r֐QqRZh˂m]emn<$$p.4 fU"Іw|RT291*խе [K੣-H)9݇rO##W2J(u,PJ<==dZ%Rj'BPs^:65v2cHZ"LKK8a4"UHmhd^cחwù5 N^lmE^SW.e72 ZIp69z)dPzpMǵ>(+]-osPwSoPVmі1FYHǗ]0=dd1qlQO8н{KvҸ~7FGnK2`VK/IUdlےmS:K. p^wtCfjo |/@0+uXh3(aϡDAFE"*mgW]3) 99@ֿ"TpJI(wym1-ˆpfs7 :o)2[{% ՜v_<ۼ54V5[yx!I΢']i5}҆H<` kI}QG+Ƹ|qZ\+>2#>넒8׆q2:G e9 <0 X_J=t{qY_ӵͫkY"K⌱v4bt|tuazE<{  ;VZ f{uWdSI̴;oS|w ~{:Er_⍣o~4ؑN'//{TDw{_Vv[f>:t>RIiX5h:ɗ'+ _6 W_2YV$Va78YJV%3>dՁi tiV˫ޞdD9-d]`<:YDyjb!T  a {=iF{K_fQ)QSVSs.%߲YQA >tL7タ . 嚌Va4Pޤ~ >~^N-9/87o -{[9f }W)hLJctz>Sh$Xϵlkrv"fǴsΤIB r5qwf2Զ]yfn<[~i|7LQxufhPwoKc ,iɸGv=/W  `۷עii|Ø!mxTR 7Xc%b[vL -3Q̚/iV%᳐1RL bf̎R HIFŕl5nY;uP'7X)r?.)r?x*&JeXjй)NN@G,zQ:Pu@ZRo\Kq_:! V^qqM$Z$"3㨩6̉󬦎ցu@OMBaAY;Vʆx'ɂOA |!.F{eI6яw P˩i0)(հn@w>tttt>īE͕VZ[1OvxVT`I'y7.ϱ𹛬\Nvox;u-"B'C"A OMz$TD(r`h/e\hˠSu-cvDKQ-LI:F02F xL-q HcmXz:WQyܱŜm1F2 (ȸKdiϳ 8(W#vlBJ8۝W0A .FYv׎3(m&FI&N5 \Pȁzha2ϴ_[Id]I"JYWȺ~5[T%8֒ޗLN4Br&%-:tlIʫ7b@>g[QR v]nFP³ܘc=v.]^t%zѕE׏^TFȵ~$Ky"\euj0l7ahI4d﹆"+]尳vuԮ[ZP${Gãi;h(锻߆fء$D@q^=2RNDH?1Q"k,xDTziQr >r2>*:Z%]j9!33T^eG_q͎3nwo N-NJ8H4_&LdwS'D7']M;ÇN觼.Mj?nڿ7OmcAWҞc%j m 7I1ƓR 䞭R׼ЋT殤2J*sOe9ɿM'QV<`ȷ 2+p]ɆJ6\φ5F g썡 y+O1e "踚Zl62OBXDJ?AvJ:@ X3QCBR$,ECsTd s1MTҞrgv\m;EAtzuq^.w[eS_Re~,XnQlM/>lcp3d|1zw=o3zę2gr$w_dnXϋt2.$*Mp3u2}^VXx#S$ zBA$ rtd9f˽ RH2 ͹}@~qUGh3}^T2Fj2A: IWM=")g0\%f4H}QUz4*/59(}fDEGY"T^FK$zd]JT/e!#g_%f4H[q/dT^iϲK61}J"ĐH&; FRT+]-,2TxD@`c `!]m9'צ238n?B \OPKPCGD }zJ?%+ J2>AGm1>;ʷjZjjqVKC wri 0qC$ܝ86R]Rd[lJQD K.؁dh(8;SPE^nKQnԬe)p!) !AKQea?%U7W +9ҼY.YJt&AIz:eƍÄL ۿ/dVFj56 r-{%.dRZݯ{#y~};U%j$~y]2զ&l^lvs+S{g+cF/Œ?ȵLOCjoߟ1Wec5}kVЏFlyu OLw]n~y *ߗ6y[Wfk\^c4NG?޶ـyoueÎ*/wH~0Lu}E/ĂxFV,8yr+7ѿQ!=^V׼Y> xkd޺Sbf!υXj|XX~0Y~Ҩu]9P2h%/FU=7hTƣqRll:alfL.y2rtIF9J+%0;ֿOkm#Ehy$} >j+-$'?,]vKVۥGdv@nx4'WpJ2JI 7 B&q'8Vy{?pk[߰#OkDZø?5ÏHSs7@d:Gj!2,fc7/A,B pQ@(B!o 8 A/*ޘ/٧_{GwT+fVbdq3?{5m8qpdfaJ89]ߤ O-r4IҊIT»|quho4y$fx{_xZP-PbL)c3"[FD(8E(x7^ч0ʾkd!-M޷|sc)Ӈ-}*}/CyM[/{GpKJzqF-,R~k'()\Sfqz(Kw) :gX$<"SZ)WP=RL1uƲ 2cX"Dr{s_MaeNv^j+(Ur{/]AFZ=ಸuQ{J.,R?Dț !R`w}Sp:~ js;I_!%O̅OWjeF%>\;ubի޿KXxC|0fSRWf?t|r+vvfx,Zv0\=jVN= ݒ%|JU;`@ݨ^ ǘ5-e|/A X ~&1fV0iMqꧭz,UrsEst =TEP+?Y$g%ٔͦ[7+23˿w3䒲 w1O3`^_{qV34go;~1<\eZB(pu_%f(XAUPJ#R$+z4]_,^-[N_}6؞_}U3+TEH4*9W!(8Hso-Mͮb2}up+X}U [~IJ{^hqUˬ@䩄8.inr>_sҲB,V/Tk Qa7T T&i1N4\ES󎍙󌒉7|;䔑~иP'sG bz=ʻ+`]%^p<$;w5zp.FWyŌ_*J5$]`)i$(:.F"U"  q.aLI+kʜT0kAH=p+N5N@e.g;P|L?I ?=8jb{8OQl #ޠ+ PX؉&a1ۏ?f9"g`}[̺I}IJm@w#g8r99r[3+O,ZcW1IWUW-UԪ00ú0PaX? )6ZiZX-c‚qF9 !\XΕQDI#)iP˵f jݷͬI)*__\WwB@K >@˿LRQK8_f$ X&G_RZ]ߛOR/h),D|`j@IIݤ}HR߇~V*ZS S"^ H r'Ql&(H%߉wK@ͦN~ض/z1zL Rf\X ,-.BKط~ysIg3UߚD>gQQW<"5hWpZվ&#{$97{7H+l dn =8BReG}s6ez61I,0[Wgo*KeW:h}UreY+ҽ.SkيrK;Awt!ֈ]aXo@ei$VL PdaR79g|jibGq6MmY?9[ó4fnh1k0EMfnldOy|Y~>6gJ)]oԕ邝+(~2UdЅ⩻=0)F]d΋jk`+A<}~Wpr!\L"e퓿s(.0'_ۻ<}GÔyYdų*`׸2D0C* }%}|wҾ)IZ*<5c|5kyΩXk0֡e2/du ;.SX\ORYޏ8RU@w(56^û"k)R PD2JM)" $`ܠ@q(+9 nX0$(E7RhҏqfV_&1Fq^GF@Y XQZ B;X[XGTVH=7)`s#Z2-1-EtOĈs%z#oƋ%1RӤ"FYh#M 0,ng`|>aB8#Q|xL+-Bzq@w+GtTS)Bbj)8:oS A2@QA$T<(I@kc4daGĒaHbϢNrZJjZ-ecCʂԹ;D &b}-3Laa$(Lm)N 9`K9GuHR vFV=:8 &ņ V($g*D*TAPEPR^k$^[`3/@id L5" {+dSWG1:/"HYZ\/Itn )/c0I RSU0bQ*MSI0M\JodG}H+/uQ iAoQ܍\ g~RW( Cb6b;#K8B{YXb^YdȰPq% KB2!wHט#,!ߑI*pϣQv>EZnЯ]U 6#E$]@ueHV u^hS*XYTJi2lj`c`sJ1TCS6bnLHO4m5 mpŢ: #g  b@6!zODbWCSy 5 [Aic(#2R"WJjyps!]32Vɔ.^ٯGꐩ O\w?yK;㜓oțכ'bǗ}y " D[L0 _m6/Zᷛd|u~g&fڦt ۛ+&cn_D3.w]L(n[I& R0$tLv6qVa:<0A TZvC4gmID/4EӫCx)yP4Qp-\EQ< 7#+ D 8)-p91ƛ6(2˥cU֭DBN}}ֵZ@DWVWeX4;t]UoVufՊ! &km8KCUZ81,ŁsۥcRm(\I^Wwu՞ DS$>?E#jut;ެ4Ej%L<|jː$6>)F@9͆^[rAɚgZDpňȀv=]Zflb7oz+ \fגDk>5._ORY譮]%lݠ˒VS!՟Ӓ +t9|,{ vY9dqѵC7w\{o&UvU2I4T:"zDFbtgrC)yޙ=aϽlʘrVm3`{lu`!ɞrvz ms{cxu R|^ < ;.ll 7T!]{q52&۔U܄K\U)U`o̔T $䈘.g)22Da%E$>/$`[0b:+|* τ)TQX4W.- *Gd)B_:WRwY,fg}3ϳqwԐ|kf|jusO|]]^vD}; ycɂ&7NU-0LɵCnՊܽViUj濮ZgCbݺ"Yb~JAFFZY"K$hV+=ӀMW /y;Q&l:ـ =u'm̴ VA=BؓУ@K@>˭~F}O/i3ݯl3̾E4gP,ǛcfN=o. G~?=qx]Pe$ r*ٜf:_znFٗ7o}<œz cJۻŒ6=]ч+]ԁ8DiELTHvc%(nYOi& E6WHDkgyF(})\HE}D7ިoh4:I$Aب()-Z#+\euX@֔* P^ws0W%:G&A@}H9v;١6—52B̪ .׫A-6> CpJb=oIk]&t|*:=Y7VyP'Bf*pNC]꤮/ N- V #CQwefo񍛲 KGKMh!P9 s}I35D@)ph(nQ="PXA/ 1GbrOgl>@nk7/:@ޢf/ SDWdjQnMyIBu.6uAu(x0* Q>_3FO[v7Y|ZL9Im΃|zK<[3ì8^VY\}zt./d Ubmy}':ڊl@0ZeaR#S!mHPE*AH" 9)+sl[msJ3 s-b655ϔa6"z(qʍ>b&o'M iiUq*5`©MoY 濛4ꃓuPG?W_xջ %&S\43EY%?*=JH)Wo1糗McF<|gbPWֺ6bSiy@UK]'OK 6f7̗-KHdNk;iDÝϛEQ !ܻmFVQ+WWH@S_ѯ"@>p}1%"cBSSO4jJ1hXނb]cD6j^9RP԰~FV2a*g`ٖĀ)pv}|ZT;TZu_p`R)j qp3cLGv1}zSA_>塮C:䡮Cz3z@6%:3yV*{J۬sչ4ʣzb=˩05)zd4*\Wy=닮&UR;Q͋<-ݢzd؝8;kz-F:(r;-G)h=5rЖdSBI*It6%_]/ho)-Bi' n"wXKjVrĶm"GYp8ֹmiQۡUH2MHJ‹yvA1(An kW 4ՐL&Y]kQ&{t*SGڷjuP~㜄&l3iFuuBpMl; zSs*+%!֥.\ٓFH=c[LgNQ:2m:u b;7sl,]l ?|M26Q֙xlbg&mN&' yuj2om3:xlr8+U8\6fD ̏Żׂ.!AK-drrq4!:C5$yo3Jh2_bQѶ]&9spjV%ΧzWl_(==*zgv0pC i_Q !kN>a=0W=ƫ]0Zw"%Z6#Z~.$h)0_;{R7;]n2z4;ImǔݨGGK~BØH/7+ )*ѯƱ\ )[!NѶӵBxLaw"]9|gUUrq+W0 #?7f >_}kКW4n˫uQR'Tȝy+B*m31JP\ ~mWix~iX/~M'{)ü=_Ö-fCoE&;c64ѨR^Eڞm bquyN+./&LdWO!翰 B@P2Iu6u>PBk(E>Lk4D?H(})9TGkJ tHtED37b3Ԟ$ZxfU@U\F⸞mT1,Ty )Uaɋ/Ѩ0FmArpKBh3$!l.ea F\ڋ?/wEukRV-M4m,"3Q \wh %Pzq۲t9I}FVS0n g: "]o7W ڽcE- v?8$'Vm~ɑ,,r;3C;ղP7}y}VY~/ vx+%adcz@*< ,yӪ.[-( G=I IAqXx`̓N!@5#LGକlD~f1_%}:P h2:&SLs mHWYS* Q3H1ޠ4w>b[\&%r!랔ȏJI"  K  %".=~d mRN_:ZE\Kv-シ6lF6% bZIV Ra ڞ^Sd5|3̛VK=<\f 8tp W]+7Oi\l$ NSĵצI"k0Ǿj0[0pJ~MHwʯ yfd1>57H1ܯ bsSfD`zP ףW5lY܎!q}+(O^?h'sOjgE#7 zx7fȈ16Oi qSCq80%¶Gz3"vȡWucLsr׾ލo,nEOW6;P\qu47!'Q(P RV.hAJ Er #$`CQ+ I40hh!>k_F,MJKNRBǂOxQ)~.ԗ!~<$R#a$W2wJ wy H+g41hm$DB,ju~T!W2-T^X"f"[IaPBa,xPI%@r3/#f-1:|ȨR̗05m+0=RjmD60bmp? +MV2c?&bq h֬9eEo+\ŶlxQ]^_o}o޳pZvn\0\zԜyw^Sx.W2v<ȶa揵U BC[VKXbjK#> K$j]-sʶ׀\*9gQy&|U>,"jE]Gv IK *d9YrȬ\ mSz0%%ɕ57 1CF U m2 PhYZ-nK{Mܩ~"~*;Sb|ՌR37gLrdRYlA(8 ȡ{X!,#ϋBAU=#iSx.9rmW\!v, ]uAx3ՁG}H;I 4ivVI|< xB, ƞ[JѬ3˯T3; N-aʩ(Hv x*Np! VD2uLEEYC.g]X:&G$2ȑcQ]FmrDMEL1@0cG3i PR5u?JS~sfuv:Āg߽oj _Gmqb8Xeux N71iX3lfZc޶ csf"c脙궥g]mYÍvC?lG&gbCz_f@ސhpoӃCxzzCBڮMOvu@Qpw'T>UR\# 71ecȦ bK!LwTN-\C1DҥW&$ szWY|2=A(KttCLX6 z)9ai@!W7hcZ&W7 Nc8 iw ֨%WdӃxz'Wu{s =ad,r}W& 1x2+%Op ktk薪ԟ>TG@+3/uU,LH9׾NK^WnΤޟC+{-om,W\!ȯï!Uzr{A.}nON".~SLg .ZWX7}k7oPƮ$E u :Ew ,-Eחwzd{Q1 "EF/[F^Vhm*I!D" 9Pb 1+Pbz@Y?#j-"8s-2vOs F(X:4F%S) h-?9]?UzM09Nk"h2IHp&=+(АcNh`9*skQD"g Z &:l&G ]tjOCڙna wG5۷uͣdnCQj|(7Tsz@1B\3'~d[`"Na`u[el43pcoaxO?G׳oۺG#,~ǂee?KSu_Vu8 (I8׊@C!=p$$p,&@RFP]I IIˋ{td`N@§IŨ1\"@Ic`ӎ^R^1VKJK`>"BRk;4;ƪZ+4VE %2Xq”iSE$$Ychj)hjNp. 5& xΜa`;>Q;֝Е-Vok/[)/oM?E.\VY_JOKJh' po-91hγ($rP1qNOMUZv{U.Z,|镎pomD̚әzжfmV/cd(e$12/MC.-F(=T͋ 4In(3IHAv7j-s=ʚO6RX|Q{P@7K0fbI00 gOJZ+h!,9cR!Fh 0ghd]iQwS kIrxg ÉcJzrWݚYYi.3_:sV% yfJ?ok =vM&4Yߑ lc᬴֩7K*,/>>#A,?;Ԕ1ڱecVU CM[3t)Qu!TPɮ۫%.Q^-pՒNrGqupHo=^k !>$ }lCƜ4*L0kq؍2es\ݺ_ĐyAK3A41iK2Ynev@lڮV0avIcz>k[uԆ01rKN͌e[lYƝ2:۾Pջ(}'gFJ=,c@ŵw덐<8A~.>;SKS{?`f"6}68ޚе錁4t5M =a&)BSُ*gg7@`  lޛ 4wSx!>cpQ*ٮY+0֛ X OD-WjyބfB# ǓFAS8_3l )ƮQMAwLhh9O=m ?)`zf 3:޶ud2wn#hR!O A]XfS{Q S(07'R4̰(UFKu?˷T x*U&mTpn3e8MDƆ}!bp)7c*9ԚCETN5 px+ 1!(A+/$1S_E Ӊy*ӭX _ǹ,ߠߟ毸rexoݴ/@poX_Βl8JS6} {;yg9Ã{OpEaH3y"\;l|' D!FU#H l ;)Wn0:uGZB%Ȑ;~Iw8nֻ,&)nS45+?_L⁣Q2 0|S!P pJf- v1?do Ȧ>x!I6^S_#S:9 |s* c/L|; _ ȭI+6AMƜebh#t٘,] ԏOR`80ًd?Yѩ Z5 `㊷:/RlMr%t:{σ{"a I6wH0(yXA1 [ ) AzvIs@BeI(t$ '_`J &'pvGW=mwRkP#()3ȞLO|}tE`6-^ܠ+2IsuUC&@דZC2V4^)RB=Y>{v J]( Wl^ȳ&㱳Ӿ@I b ߟBB9 =743.BxuXn@X~}"`v3MTޤ}19MIqM S!xol! Y/$ۘCg^9)sRQ.B^4NٗE`Yͣջ$6T)z؝8XR}9nU5zIJ3!j{Xf!PB~ ]ҭϗvsu ԹG (@vAANJ1?&aA6.:=N!{K7R?Rt+ 7Eu`)^ i֝mYbQ#3|a<|6c_Qޣi_n]FN.#G ܯnm})PɞWh\7##Nɋ(Bm'>z;엍N23 o Sx9xJE ByصaS A{/߭͜"X2=J!d}S:v쥲h1o7GIVt4vcO=M\jdw~s?~܁܁܁AYp/ГḇESʙ<if,#\JH6)ʴ4SPAw ErrI652r:Er7J O߅"?6z,3q<}H&x5`)XDlVZWW{UjUz$XXԸTC,B X&(fRŀ4o5N=k1+H* e;j,N |C̶S*?QPɣCvF5ҧq kJ|Nr=3(OOL3 Adf6!` I(5l>&$|=9xS|irM0ÈpvwSk_w.~Ac(PxbU&Uzэd8;$-+NpaV.7ש" W [>ܿW7U֓RRɹR [fjM%n"wh̢ZIqդvރf-kR>=>TwVb1Knd_l:~$}ZK1HU2lEk'zxX _ѓ{2 - =Nǩw Aj;NVџì;6`oӝ}J\Dǡ#ػpY|/ܗ:&Hpv56/T$=:f_A uq %V1@R^DYf9c&&q$$K-FSC@OUY&W˾q`Ȭ9P2O:W$%v6H$c۪emq )d4AƬZx4}7؛$3|QI+^vM.g?ycO#< -K -]G/u]HlQt_nvVtv8{G+ȓ+kR0ײVin)Nj9H+'[$Y "IBsm.S:4ZDv 21?߼za (;bl XT!sڗw׵!2~>49?8q%o"w3)W7̬c2Wo$Q;笱E5fr?nkQoRgvYsdi9Ep`J~ip1"MM&00*ѫ1J\yARԔcjGDFN6B(q_+q;ĕ懆ʉ+ӝ}'p%gO0[8D}2Ώq5]x0{pʒ)rm䮫TO#JIH m9J&[% @[2QI HITV*R+i!8$` e M :s<^lsn E:jdS''{9?V_S:9 |sϓJg jvsWyC%R~ъx}?'1-UF6?}h:%j ǧAv`ogòdx pJu~ʃ{{_nL&\y{kNNwn'[)iĽ62áeK'L0ڡ%L$`U#J6j<"f2 0$xRBVo/V\xb1s+i>>c)`6?TYc]q bvv85Uy>6{Eoa`%/Cn!GDt=CHY_Cˡtiό}A>?&(.;b.aV2[y kYƗDOKtUÎln DwB#^ծ-N xtoRc ILjЭIǿ0rߏ"L9dR= ;,=Jg*wxW])߸@[w~d huc:2wTTO4'Zqr-!Ve pѫ;ޕ&@XO56}G !yN4KU c|"Gl,m%պc״`xІh$.yy˹ǻ 9FmTO7 ELJhg8y"}Vq{}WÌIWIyGy8{CɱwcTE }.zV?)<9C)LV Z 82,oldtkdd`\t =Ct:IAXVn'?yK@ɧ ȜNC!-wj=M0s7w !ՉsF b(gD=4 fSm[~B4VLrML:E%LEd2#kRia7J?egxFô)\*g?O~Zrt?E$d{iVٝ_܏_7ߠ~ zPޠ/Гḇ! } 4HbJ12K ccZ<T 86S%w ErMVJ$]$`P^7m(i77פToW];!s2Evc/ {㇡oznr]cv݄.$PCW[-{Uwp4Z]lܧ$6q F/Qi%# qEJp mTYY}Zžm9RC4 RSi-P!!S)TXHdLa" ;+{w iq)v~M4 hyAY%Ddr[;g1?SLdTί-o;?bfKY*3+ ܑ*ЎD!'IN0vB!"nBP$J Ś-C1|Z4 ;L$ݑ1tR@$N9`*%0)3 br#Zسq<|Mwp?#0b'j( ڝ.2p*tJF F q*PPԍV0Y)b*muC&@hm-DJq=ԘMh|s@oY!@1ԜgZ8‡\\G~$WQ%SCR̊(E== a.R$ܙc{fw% R^FB(PZjйQu; T(BCBLr!/Q-jdUs%b$jO. gaiF^YM&O| ;6Me!I;6Ќ%)$L(yP6NX)~^rz(_j1+}0dj?tvXBTqyQ`o(q3ؼs-aKv>PSX.jdQ/G=ʍ7I 5Z_aJJd=;)Ol"[%{_!{ru3' U+/UCg,9u!q⠪ꌔ|;N/ܴS-\)NqN99c:Ua.:`%cY1dETOa 30y*rʑ] L]遼婪2vi::UQ.a{IH~OsW+0ucۃ겏_V-;hh[ .ДSTTvtdTng?ͧI~./=2>/{ /zu[obN7OVsߛ_<.k2B 1CG1z?깱v{Qo@OBӼ»~Λ4$S&WVXA4?&4ͪ߆a7G0M_&3s,ү@cϤ`{3@~:b`102N/?GkOԇ 2TtH,cҒ߳ΞK3tx9bU+@~ %5aٿvg~U_Ǔ]gf^߃|[ºsoYO0W8Wy~`G۵_~ހ3Z̼>Jֈ>_ћY=gYYyx_AH [Dro:ŭ!\tJD O4L*"), ̰TiƭVc} 8Q^Pνp dFl_=ٺ3z/dz5's\(`sX4wAD li&IZ?SLD1`L-yl\F=(_c(_%S5멛}U{s y XXRXy(!)]׶xUf׬&I IՓgdJL#<]&t*LZBᎦ0+& Dx k%V$pL&'’epV$Wթ܏zFU`-;ذ %51զ5qgWGb3,XU Uo1[vTjl aK}lvxqsqT ,JA۬D5 0TInPXXb1 l2SJCs$G0 Xgɲ嫩7(Bp}a+SOV ]n0xmn%V`-/c֕,g[5łpdG)tJdbE'qwNc;F&7Ք/ιkN!)dX^c qTcKE:H.}vOV;OaOPL^P~2zv*RvK1˜.X/=Z(wJ83p=r^%b<2Ф,&+]z8Ud"9 sl/\FמY -+=M/ XAɀ}x!gևE8 0&@8yjsˏԪX!:v'h*3VSDpJT)ey*,#wRۛJI-ĺe$R L1O<\jݽ0ryCПz|4&`J,N|_6J =cN+-{#!a۴>/t~/`*l4Ϟ8eѓƦ'6*5=y:Af>ǾESD[$`g͐2:wudY,f:MUbfM}%P1ҪmL$*7V؋ځQy ;JM F0(Hhm;ج$$Eb5nlSx<£\.k6 #PޏԢgYieڏutYK>2A`ʽdbk6bmRT[0Ujwu˨Y455 @IqOǵOz}( Vdl}/lrg6|3|tl3t5q>ݤ X/o{͓a # v(M"(bӚyi6ޝl F0C Oih.)K\z' E4Jo}`m`b1(#:cTn-в֭ E4F8ŭIvb1(#:cTnřSU؁9u&uCBr)ŎuÊ6XX ʈNU[sSEfݢ -kꐐ\DcdJ8w[8Aщuk0o͈ݢ -kꐐ\DdJX7juŠĎQź5<l-в֭ EhOX򋃋Ή N&Yz6>\X?ˮV8Q4N$^PZM,iaB8eelIP90  Lk0ŐIʩ#9…+9šNGUbFI讽MgXm;߲3d1D+\Q5Dw mjMJvkdEyfJTv1C1C%w1C1C&I UX]p16+%UP k w4l+× G Mv\ 7$anpeDǑSXta᝝ס܇l&PeOﶉ"=Ch{Lͧ CyN7 (/5PA`1#+ ŦC,RqX0dc56CX_{q{@Ta$lưgJT/5P %<X.im'Ž8|RHmS.}[kxlm4#H\EO;f0l]6]Va}]t+K/A:Sp4ƨb.Y $KV7$e-Hy˅RJ%PQ,6EaRԁoeiJ5iB#@JH~s}UbvF1ыǼ)\F| Ӗ'SHδ9Q]vKFu * 8¢)jzۚ.2TKaFT.xe;"3%uw?l}?L.*@)`5!/o0E Pi;;>u3\B iϐ y[c3a1& Tj}!HN_P) .2Yb9&c55S@Y8 ⩧T%i":mI~%T ,c8B OR눑 Ic1q)ji"+HRx(ǭPBje"[py`Z ! (N[J§2_F J! $)Lb0'*T&4QJJp@Fyl>M7^45E\3T&ΉlMy,a]`UeHxʼbxs࿃ژ)4N:H*g^,2 w%rחᦨQ F:Vh֛cV?C-Zg_=EAC E{m^9g ~˃<S<6[=fqZtUbLWj PCLF{dxii:HMt`6F曌؇ qorwA2yu=ӷ~,dZ-t(58uǚoo /{EՙZLRY?{֤8 QOgS_&NtNL\:3alC K+ ٖݙ~lcRT^$s.[]04xyz"%K_tKFl=+MMM|9?Tu90Kr5rǢTŦ,,*ٖ 4>Gd7f7Q z6S}N jRyPaY' jBLj2TK@i:!\TUӔ4V,T %e2̝ǚ*[O(P_z <նXdc8٬(0Y,3:[JH],P (NVk}t`.\as;Pw#yU [mxT(a:S 1dcg{H!ܱ%m1q'l~U<Hp)+,$#E,nUy \zi,!մk[I@I,˓S EQDk~JW`Q`_jf|;X !Fn?wZN/,!"\^dR 8uCF[rm|ȫЗ8\JOkH}~ TK5}"N}k<9BH\&-VQYR^}V^N)!09jzPőL,\ZV^]x";1` \@k^@ 1u[Ӕb(jm",0B]?k<-ϴx#УXkOFa'/ʎeGgUThjsg\Qʯ{vk?i#}w֩_^P׫ m'ʀ :Ce@I4м.zQ# i d쁝:mJ?.s&$N=h?+`vI p!7{9~2{D5BctGw~3#f j:w^怱.2%pʡhЃ I8٦NhC G[yX͓cVm^V+ܝ(]Y}g0]_芑ςs$Jv8 N[R9ZVp.SOa@'b(5$oTpe^z@&0'-qO+Qh65NYvސPh\VlC\Aۗ<]zT5԰Ec8uqnE|,zSȿt'1vsӂ,)jSwtpFZKw8 Fp\r@7K3meNЗ5ABG >@֢g.~ٵݘ.~ݴ0_}+0U׏Y/TIK^N"ǦLz6Pm;e] Y;oj$X2Gy]6U}VպՒ\c0i-7[ͻ[ n%#3P+E{FI7tE^ gƥ`NLCK@AS[r73@ Rw|s$aІ=p( (jTfGfDQ=n~Oc@0Wq[qFV}V]@"]ԺJs %Mdκ'rF K(r6,Uro\B9,l 1'fs*v$1N@w[?6y lrAǐDg :Aqz(n%BGukosg1}A$"T^'hOYW2RzѣOѺ;*-Nnfo1 w %)14]vÍIJc942"S\/Lg7Ix}HryRJǶh3|EH)連BMn[sĕВ)%Q<=-,FK@7if,_ТD~*dYl; ,>,F19zN72͍i">X X{wUXM(@L̜(%vvcW⾛e!yCQ. j`Cմji &pN1BU+IͤJ֘*&ɟ} a4a1KX"x0M,09=IC;>J#&~ч<@ /C۪NWuQL׹8: `| xڄ]h1Tբ_/!$.0/2bkDdЗi4A! 2gXUxJoǢ>1$6d'tQ (Cxܳ^ge9idX*u.֖ x' t 4 gڻZ rv:b:ۚIA@ݱC"+ Zgi֓ bߥ/DQxQ&$[@ !4.$ƒ{H<1. 1sfiFt7Q)o_`R?&S QwBYе9҂EpX["{< nC}J$.rttݝ8&#8%a4Aʹ͉†{ NfWtSy!D]aD&X Qҏ9ACcesrmnbo%S dfy.KCp]TT4Tβ";Phy,(ЭɱP*lܑJ[=z:|DyXԥUV (NVTy!8ǿ'g8NʪNiZd»ӠgfhS'鋜|NS oea& \`i8˔iF)gv6b%z V KʭTkA7C8Ǝz],K&0v_higa\?7H9zZVF[PPD}cCrgT(<+P N>)Cu"\4EH-;5tr̉Ĕ9oS 6!+~li++c;At> DI9ԨͨBZ*|9+3ls[GPfJ7wonnx3/n>^Sx)=K`~ռv85KdFamya߿ojyjP x%ݼUZ @cۍ$붋6  ղf/>nҹ&蔢h(x|)&L'Ga;pOJG?&jd62]y̒r:S1cjDjueVRXJ?%߿$ѣf߲krHFU7j5܌[><&:ucuuBҬ7~U e@߽ѯF@v(-08]8}ʁ;Tկrp1sW~L& 2ioGW90ݘ},ŰRFok-QEj=')UӰꪑ4PMqWwYv>1n6^s#(|us_Qx鑯iz`yЛ>eaӛo=qT֎"n[)G |aC9Y,C 5ԯٌ c%CAtlO},%^^vb< 3q5ZWwghՃb1rV%dk)B5"]j'($cH~!i VOh,׃ %xD߻߿gl|>2]J˫i\~j榺SL4#LJ@N (0$3{г')i= %\ 3 |DCwm͍PyJTnpA VFId~Ŕ,Z@R k4Zug|½KO`kD$z".' .k;|FUt_oߍGS5;FPUwӀdavrC[/ENFydCʚ\[chaaL uqu8[O_9j:F%Ul _,^?_ڏ?|wr_^5s?NF4q0z2tr˂ic%WIT!C4/rN%aXQ' p ۃE 16 4)^4gnb Ar*,*03"`uN iLXa1OZy$ .,3Ch87yKuZwSDP !,'A&7Ό7,Q8_MZhӅ<CF"WIHy̢jh<2;3J%eRxwz*rej3wW)O6 <-qT,$cV,3i& G* 08H]Upo\lw󁙅jݷ;JhH.ϐDD?*l2-@b>vy+;}4:huUy$Q#kϤ`XLA̅R+ns+W;43*γLPReVqJɣvT|sW0 *5@r.9;$W(NHv3hkd8YuDSl%TMah)hhPgJ;9 69%goqV)>3s31iK9͔HPNik0x'iE|Xd@ql%2jmn'[4˦(L' ,*ÊLў/& )Ybu\ -y&"ZM-hlHf%YǗycCj6bQH8!-b&ZYPY?HrnwocYQLy3 V3|cX:Mv_Փ<{,_mFZ=reM6>#l1nB8v}&Pa >Iˣl寔/:J<"8cQ6e9rim6*SG9}j$:z͠On{42v)sX_'OxZV(_?˵Yjg/|uR(p{iܳ?[jcJr^y>J洫Á% hO'Gtu*L &]P8]EP_W TdWC>2Ut"F2UCQ{vKp&lڥHpZkJV߭r]B)"eBAO<)鼜H cJЎ9G)1&$uO)8;=q=dJrZ3Ь91:uT9W\iQ9 S&B3S2C\r!YM_`4fxrtX˜8jx#U^ # .Pt!ey&rJyz~rF6, }Y(EH-J/9_lXJEzO,{P ܃N@ B0} CQYi=(w2 lKpQP\&ժ&:quԛ,1;Q﨎:5エp U EYv'pP{P#kQ޾]eezs'Rq*xD[3`W XYqbuRb <0OwߣrVپ ~Y 3Y Al68*q/3 _f qA沜z['.t@u\j{=Aj/`jNk cf`9%1Y#Bo(:vd4TX5јnujR+S"Ǚtƴ&fj * )T e;w<;_^yÏEx<(ޛokMM(8Mw&??֕:ZȦ~>I(?g<;pBp1|E!3lN8Qx 2NsL}n BÞ[h H9X,S8K_XM$@h7fG>IZHvVpp \C\ r$3CX);@&قrHG0 jd^Po W0lY'=%k٩J6EQ#huG?ZޞnP pt(Ҩ3ofJc2$f G3[ѐx!rcTDaoT$W ׀!aX)pt`^00;\¤Rx1;03iB'bHi5OKzĵ˚'qs$+{fr=\ M'y >:p -Dº+6 f{|kأ\P"K+Kq͟U?mfrP|'c1Uk :A VOf:~58&Tݥ1Uk -#1+t0K[o;^Q'm44 rQGw7NX곴LzF!GNh޾z)8]9f`w8IJ,Ihb=gQ%Ld}GE[ h8#%>۱Ri'[nHWx#J|}XOa3/3SAв<H6qAB꼏LΘyfn4͓Iro#N*'ƕaM+ܩj_CQc:-6oVLߺ ,ff4 "9,$6ٌphIU ,J˘Wշ[+ND2 1DI .DFKGXڐ"QpA!=V'\nA,|(M:aKD6)2@gĠІtbJ5ڑVjy(^[&m Z|qƐzal=`Pn=&7šҶ4GJ@fH19tҶ247-s| %Y ĘZV]ƙ Vp ytDdN"!E 5#I0-,u|Yʊq${)Jr(X>$d}@ItJS/[S > SYKJnNt¿f瀕}56Vz: ̶ozpN@[>|v {%i[?y~u2㈉Q}F?z;z:j?TCO.;wy3߹Ǜ |plM8qA|g7 7^+ִ.E2W;J2\'mH{-%MqytBIۼ M$PPouQC-v3u ~3ЈPzRGi#7V12cG:@J! jHzS:SfE$.&#<.e(vu '`@l]Wo,`o{Vhe']^xu}F5lۀגڳq:@AR(%؜Eݎlt V.ppɏ^Pfh&@+FAHQ ʬYh^!cELY) #f BʨʿɆZ\œ(ya,oHaUH3Ekjznum="%cl/~QW~]fp"|+u W7}% љ2ֳ4L+ {?o0ǬkGt>[(mH?!Mz*'W洜4C*^fn28S;݌3;Gb[js z`KA3&z:2d:K-^،);FaKl,i9<Ț ЂeHn%/Y= ,QkrJ6[mƔ#-c}قG3[j["fK/̘}2 ntbO7O>҇ԇZ%f&za6^#]<9RPrWR!}@y}bs;6Z:cw/iu{_BWYANW6GA|Ω@rW񚀺)[}|p5jpRLu_q`λGiՈV)diIJlL|dR#`V5Sdᦦ?/W#b kѨS;h |t#yG??o)6!lCmq5^݀:7,L֚gu! 3R ؙv!f9p;n˜#9zϓh!!ERxv,Ѣ%0n"J"ԉ kqb0eAVSlvYITM鸝;a$CaOI+=2h8̳>F{J; %G@CP.ix:S )ߞ%W|rrwQ.-K779>Kh~ȗ=#c7YqX+n<|@ov;1䛬 Z鞑ov1䛬yh<iZ $q+ Fi/` <)_i@rR sH6%gZxj|1rktXt&'+3Q[YcX ld. g+ z]Nztig .O#]\T "u9V; -mnCl,.xkCcͲǏ6 (ܺ3rGϟq{yda BJ>Z&Z{YTTu7?0g᱅7d ?b lZO:60֎VPWi8_)I{` ɾ^;$Yܸ|И,R&X^o|-p@.e x(|9OWVKЀEb sXϙE C<4h/!ITx`8ʹU ?F+-yc%~Kc]_(frjZ4lߕ U*L۽_|ӲrZ^N{/T\"Ned"`0&*&]&2"dOlB7,_NW>|siyoߏIKUS ĺ>ɔLfzu3=!G5ci:zs| \1oSrgOı;L'*v$r8JFy/㗓.DA?T*-dxKʼd*4K%*a,Hr^8AaF 'ۑ^LavcgR/E@d()#y,l8Q((b7 4xn)\Y'IFfuf,kc;ج3$WUu4U7ެ+9br Fsx0QP\//WM?6=mWIɆbG26ەr9)Em+0rɼrZB{ڮX%ŢJk@IގӛH>'\(oN>Չ-H.pxqĥH` Ho:SsрaCbRc(Fr ' V'\`2[LMk:lt;;e\b2Zfmmzz®Er?  V5,-x#.Q<簪4(| P͖Ge/%mL{.i.~YQ.b79"S>赃]Fե6Q0d״mAn"Ӹy|?][o#ǎ+Ƽ"m׍u 6b=Hp9Ov${rYijZRˆ V7"Y,rLŏ;T.6gb~;iMlKsOuou{{}_qpyɛ0&\ x̿][X wY'J"{aV=pR G ģO: Ь>p@95.q~ FJ[/tў~=[FÇΪΌT` 19%h"Mh&12zOuK*t r%5QI2Mur%)1i$NPҫA aZqRvq_/e7 ޑJީmjXגNPE6zq5.#,1"a[[w fc>n@Ҭz{Cb9+QJiPv'Uw& zKy>{_ތg3Zk/V_'?]n7o_/E-J}: + \hxJ`+\oF.i1 OQẺb[}y &ǖ<-CPV\=~S%BҜ{Er^a0 9vIAx1Ş^Iw * 1zQCJ>GL{tFYG:Ipq J2tf5 6YfppCNAv=Y)*y} @p82ooMxFȇC֏J-߽o.͝nNg7۟IY^xʁ%9~';;pgGe,+&"|)a28ɮ "Ї#?#MJ3{fn?Zq03*_~4oשs2ueܢ_1˟VU],cP׷Y~?D 󫏸8ST_ʁ bh=6w7 ݝ3`~7mW{7B%u]_b~g5"`~dy p OxE-0R-i'2AK 8|@5̃pq5ŅB)xI2ʨ'=FL*LFRd,:99$("<&?OUZMVùG\`aId YA1nZ:|魀hǚlw+EW/k_LJ h4xЗeȾ#2 )Ag^ S[I>Bj䝗v\\/Kg|}=rV80v6b1!w#X,O5Iۏ1<(5Y4Fs4: Sݖǒ;l~Tsr|X6MqȒNiNSԅݦم=0GMb^Wm.]Re; _ mw uvzw >m^^wg8Y9?іP%gDZe:pn7} p)L#Ah6ceX]NXAS9IMB}M9Q,9d1EJ rlw)1 QAǜ"-u+`4qP 1NJo1 Z2 vVzrVGQPa)CP3v X\zYPx. ..-TX1J ⲣP/ːBAh/'ݷb"VJ'jA1v\ Xs~=gb!k B\ 4oլAa'4dE JPAEy-TP @w^B D,9s\U~Ȩ*Q($4"A% %*xRWޙBև]E(P&uN#TwٷsУo4@e >Ⴀ $5F)Ŀ&/R'?GQrz7op8ɰZ9V&Jht`(3ʾ@Wqql'[I(!-Po* @7^$bOIj+ @udJ0F',!Aqx9wTΑ1Z$2iDD"hIWl)LȘ6M0dq%K< .H J#M 3 0@Ͳ/XF5ׂWKMP+$V|p&mL[%PiuL:-h&4TPm$Du VSeK I/b@%1Dg0 £$QQxVZU(׵F\pW2D[}NQ\|`9塀ձλOA.?<guLA+`QT^&\}0 dQ.|`־{,W~Xr`fb!Pc#4&胤$)M3ޣ)U(y&]3m.h2IF]IrW$ {~j3bC 0|2*Ty x5 R%Y*h.9G"1FĈn"#DBcܘH4/V,HƇP;K&_Yj, Y^X"#'^l2wuUd2B3Jʶ`YH<ՒP1vm2*P"Ղ7&hhb*ے;lExa 1$6z]j>/'gk>䅠3=.A A~ uyˆk"x!jth8eV " ?i{qCa!<㉕Ȼ!yN=H(.)!ϕ`_ + gݔ ɼ5LEJXKɬEi0)gXBp^l8xh1.~`*c ΁HZy eyT‰9(T $ 9F*}K#G5.R/}QGMI͒0A$Rtp%WmjkSIuTTlL83S1F0  gc⌜=sgCQ3t(kdRj0>(j ^8/4$aճ6[w{í!G9DG#G tZQDe!50[vg᪳8rg_9}({pyOqIFnё1]H-uG3:( {Fկq@$ k,T;+#AT{#ATG#A``*꣣q d>0 ^d>9TX=6)An뤡ZU_Rj ?C3ǯ3o&WVǟ4WUlMPDs(bs^TQ+$ 2y6*#cs:l9f 0'転.!(7Nh=%C4jθ}TUsW5r^r0hKx֏4ZX-sZ~զq=*f&%TvIUhZܥH{X D3@»1(F@71 5bUf7bw.6F5St(+7o-SkDMbRX3"e{J, A$GJkhݙWs@9kh}K,ѺWO%'uD'V|lN5cL'!:9MhN&Leva`8@=\%0ǰȌĎ˞ޠiiF1 C 9Mya2afbP_:G'chiWgښ۸_aiSJ_\]ۛJm6vI^NR*\%&Ŏwki )jH ¡C'X8h|_ 4zr'ݘ;U4MtMJEDiT>ηKe"`Y ֹCf ϊ8X~̙Ʒ`Bk6%]au*Eb\*ƥšns7 ^7bo> m5]*hYQ[.xWa_z%Ft1k,G3w[yN .4Bn pIX痛-&{qf+luiso6cy: {zrH4p`&գ; 0i^dS g>=^x߈N ]>6&~4fsLץovK0藌.7en}~Կ"=tn֬`c.Ԉ_аKaWypM9ԋgwb3aϞi n b)RB`Ӑ3MYr:?wi eqٝX}qԘ|xAVU~f'^Y r%3:t`pYGt8A/Ս3-NmmBWJ78f982NrMZg-Q+]pG/O4q&t|b]s?N:Aë]qCe6W0f97v wrٮu68A˓3_HvU`TB2ً&ǡsq.ZcNV Z'kQtqhfsSM^gw<Jz qSGсq 1:Ir &eW4Z$Y-n|X )H+JJU#gNbxHBtl+wl([\7&OuB9WmOFdˑfɂgŃˍg'o{O iŹ*U!$WڮcR%0˝bO rsJ>jzpTKY3gwsT4jb)ƨhs>}\5(P wl V&:=׉s ` ?-2cp@//bnP{﫜d+#5y|.ӶtAD=k|1%o?{~Q=+ф9`sݰcDy:~%iz\=,)5&w.Y T{Y"T9"{(ԝx^+W$Hě::Fuψ2\x"r|E$(8{@/C9{)3B2.!OCojF>F̨g8@Boz7 zЦZS6w3lY|~իدg8 g~=HlQF34CFYM10m#p4BY9⭒wc8{ 8\*(ɴ<YL?|+0=rz[\b}8@׏ߘ+# 1N]'1R)+Ht8j!&e(H0J ?)jS Bsos{KffLK{wJ}տ>|z@)Nn`=(XN0!0nA.3_$MB ^ DInqz~]DR+]:/Q*:;R*L\}3lcjM[XWk6+sV6ߛ>'.\w^_)Q=XV(j8/odeee+qvkNKh'L?'G \u4r9QURgkAA7#{S:ʰe1XPXd VxI"rN1YP`[eeuc1aR xbg"!.eqa2\eÞN28Jq{sݻe /{/.#H/@$x⨐-TF+w +OJK*|  / 6G4J0:pLh`IAXBb0Jʊj@k'6|_S4N)X_}sMOSpJOqL5OCY<_ߙY׳qlEN\/[|q podws%hH#s #^W?OD-"^N}?n[L^š ,z(|zx>f)!g64uy+]oƽ(/0{tHC :/ qeS\x; {'?ox Cz3̆[NrG^Ka0d5\P t:/'*~a} +l#DY!(ir1!z- C42b;.#MDiMUXd QTN&N,׽~t2Y0YϜy >4^O6Y".tm01bi/YxXu 2-^;0/9nAZ0A)) hN ʁjŘj }pZ"&j (RcZLjEz. ERѧr 9lBKA FDy +3_AjUWs޺l/l+Tj]8$nrzӱpLIn9qJ; *$֥Br LBT B=mR;5p]뺷)@qetYں`%*PʍP8(њS S1|2QAj <5 RkQke^^` B{1ye@kR8os9y94%aSGR6U<Ž tإ}0GөPf)Z̥m(Q*Q,j Ӏ^0>s:ڧe4d˛-1Oҽ11v!fM\դ!rS4UuSuºUĠT죔u;^@$Ӎ[Ձn5hȟ\EA3i[U N>X#ܛnZԺա!r=X7G\`R_)#Rsb;/Վ(򄭵#p8=B;W{ Qz!ַ JaP)70@|7>jB$bq<'+!q,tEbhz3 `'krSls]"s^B>?%.{`t:gGsJH.ؔcҢGrjWnͻگ4h1Fש'7nϿ:$]QMq,gYKjho{e0Ֆl_Q*$tl'7UVοxjɠ$)?*țutps#:JS0q -QrjZ{R6!y?WJF =+[Ҍbw4ϡBDbÄǩ%SmpRPF3Kց;Kؐ 6sBy>j+8.WT JV-c9 |6 !2cAA 03+MeHs*S)v >bJSYb7(80Z!a0,}QٚA([1ʖY!*WAFCB6b"c)(/t)(MFD{tKE=^R8'ݸ'OA}7;~liu6ZFACN4]1Z z[Hǃ˱ȢvTy$wI.U gΧY 2"C^-cZaIO|>S^}_NԞ$I1?G #y}g/I˦TL c~vi/۴c_:~& 3xad=Md}LT񷏋2-˴/⿬/J1ISMJ115qjET0$*?d3i$M~I nlXw5(dUf!?-VW~ +@9 ' @.YwU{I Jt-l F poՏWfWf>jܖ6#Ia=-# ,(&) <N}:"|Vߧn~r0{a>&p%/ίR_wbz>y;uVu*sz<"| eZO z/~#;y'd/Q $D_1i&4NWHle '[;y5|G:2s5%\= 9#ke!ܿ|6}bcgR A%"Z?y]u'cD(QG S|rv3uޠ~ore-VYAC&ЎjeGIΕVhDo:p5[+4ݕkGԪ#t{0t 9-L .aƁ#E'-awzkBC=(V?-d-\pòv-R:\ zGޚc HDpa::-q,`I+PphiePFaAe!cf# Wxz =F̈́4 f\?HtԃZacO, C5%QWgf%rS9WP tLs^!m '$},u#"jb/D[7t9\y*|L ztDEK`GUF'G`TNl1;Řo*[f'h﷚VVk"D@Rjy+`j*U@s_׵NPrrc{(jfեB.E9 V~fԔwD>SԄ$ ڊkz[tFn9AزˎDq0ݕnu-F=QNEg(WW3pR\7sfH+5W΅xxZ $%"GB͐e^tFeQ̔Whm8C"\; TF&D&A,&84z`M %ӏ[ef$ļ wh‹?0=IA~0` 7j ZcN6#'B$=p;' *N@GYrp _BV)ocJ P*ҚBZk:GKVޮb@*b!ܣw/ }\@frkұ\[3~+wk6|>|%Ie7z ~6D(2n"DS@qDVR=nQnܟBY`ѓ2NHt5g1gmn,7FrSvRs <3H1 }22XĠB1dz%BNy2~K{2%U=_'ƠUJ}E,Ņs\h$^+M CH  "C8`pQRYO^0騘K))EYJ)*[ !U\ig! Fr)( `ѿ3ePE llYQs!\[s[ŭϯ ^9u6-Rys]wYj,O8smaQݾ'!.j:)EK&.Zn&oomL#{?~x!zb/Dž ^] ^[1,TJY1l8NT~; XaZssg$z݊n8GZL!N[?#jUf3NbC.:p81H\fQ/\^(Xjkk" f"HHAEi"7_w%bŠU E'o{ dKĺ?+^WK $ wjAJ"MaPV1[Cla.j+UW"\}}tך3#]W^d1oGP-j]qbxK ThZo _5gx-TDjmuվvR^YU:1ivtv4.>h:coQq<8:SYeT9mgFyNxyJ=^_*AG!D~7lTM|{6I^zD8IC?TV}R'(X:)L׮6/Ebr1Yɼ)Uxw~ ڛPxKbguenJYc<"UZɣJ#-*LS^y¶ }pر^N͗O bsant)ϸ ;'qɵŚ r\jݒ&=#8y\b-q^Ol){_|-x) %NK<(I ?ߌ%/Ml :I PpQ7x+y.ѬT1?4@敧D6Vx=aY6rF;mtȕJ=N:)os{xI>,߼"0"j]Y,sLWGBK ~'7fĨ7;Olv߸Ǝ5Z!uBFu ) ;,ymgq4S)!%0_0':/a,ԯ~-JqEժ&?LKߡB/ҠI4ZV>re~Njtu..Ee"k"j A+!*H@9/0hZ %*$ :ޡo&?.ʚFeҁMR㪇$!Z7Ѣf0_9_ڠ_WkW<њZ7999KݪRkRϵv#QyDƘE`b$4XGpnCx3v1/P6 K4c`SZSoe/Gx<˺GX`tS;㹶82EE +${-h-x kVR(cT2}LE/E7,gC) )PQ@is0 5T(ي]:Pb~bK"x14) ?P0>;$9}\j||,.P71E[G4Xl+Hkm䜷9 1بxr޵7m$eW`ΑUN䤮)HI *H H QDLOzz@amqȬ!k BCď0r00g\4p& hIE' e2$(֝g`JHA0bXဃCL& j[α򃐇(`"D8WѢIAi( peāMo8ǫ@пs\&+rZ ?ސT1Z|U+)SOlqd-I[=Z Sw .LXxVd[R`_ʚ9YXl8 mZTx9AXیyZq/8Z$xZo;n⤸^8tGd D6 :TR?q-D\Y/lcΩ_nw8*Zzj/vx4 m}cI?p6\ƛԍ'm:7[fӟǺ>[c`Z^';fӻ#50J@(ZQȣnA#eiS~0]*vۍ.l^M.33S WQ[P#g{K h7Iҽ%%-CH-0E0tLz[̷_ _:AtUc]@VPJ 𜱠F;8Z_G<"!l@Yv&3."<++(`"a>WMԼ ]oo2Y*D ܦQuJe Cнl Ԃ:&sO4WjR[z$aEnYJ/%1Su_fB_`CPVʑi f aAd "-τ ҡ`'3)^&';;IPuuGhQ[) BSG.w(~qhB6do\?'ށ3.P`w#~XqhD) #jWUv6fi,YQTrl}"k"NL^;1.iD!, "&W(,BiwVhN1U ,\P)Odխ։?[Q`!N0WAـInZHg-#28"B֒:&O*TɹT.18LH+F 7 c~np3 1 Fsdsѳs<;;ƻcYq:G8ӡ9mq4wo&[}v~zדWo~/η3QO $~2Ӆ(N;Gݿ{w`0xƁ7?_=y'6Q^zXzɟE"zx1NLztTΖT*K+1)ZZq=IE{U1䣐{ty_cha³{#{?N^vˑNs9oO @0.bIW`U#f? L]#rxKY#w|aopQeM%Gy㒛f[r9LzZl5#;߰cE7Awp3uv?5f)_;k7竫7vnuj/qkMwx\uß0T^wq $pwK \~tOl]:=:^;G5AfnϸQjn+CIM`g)Hp|η|F$:L.]ʇBh? kqU β:w6"o3:n ܍c3(ƫFW|ƃ\-hA>L'( w&}NLW ;n$Jm):?Y9V?<UYÇ]+IW¨-]M,lWV zg@a(iڂ[ 7Fj$Ɔ%~Q$Fl \pPbR2*YB0ЋU /k %mZdk=YQ3E1_)t"P*IjEh- lI:ek%tQ_!GNZHVTXsg\+Wڤj8SqJjC^Z \ )6XO,^^.?/t܀ r HJF,]HQ/ =XJb̰tC}F٫50FC`#*rā]XZ ((%> ,:EO|UG4(r0D\3 0CٞSZ]'200 ė GXə5Xk$V;!ϺY1r-%ݵqBYNxnefL_(3wg^vJNO&`Z4xm8ˋ:t&:F*+{u ofñwX[" mꟊ@Hx@i,bYq9<HjV^ (I(׬<Qqq a@Y2']TRO5S6%M҈8iδZ.z%l^  %},Jb|!*"1Bj*b(4Ź}#!W o)ypꈡvNXH0Ys*\,LhۑL H03t[95G]Bb*igu PbIZY.']X&>/IaF!~|֤"# }l%Ǡmlϐ>{DԼ⡢͑F9=VY-2M5#D, K#N<*MӣL?vAs$!}I}EBXBHPAp[AT "X`  @4frY[E%r Lsu0vAerEPK%h&|fMܶl mdFa =ZRXRY)wFR,jAdn.c#($aF34"n:s-9Qޅ-TT6B6 K磺V"Y)pG΂4%4獫 kR\| LҒEQ}XX9k9;i.Cp\F"j8岌tK{=5ueߵKhD"D^;vy]qA[d"B.2بPIMω6X o >ZkD~w)vNދ$g'[QǻSxP ,o:~ om<'d֮٢f%QΕhݑ o48G˹<=^S>];E"$@ET/KTĔ@IJm#r/W ֙ hNrZ XE*qUF*G˥dYh3p}}qL-ؠl1cW^ ~W{6IXFF;7?<+^Qڛ8tGeC/zbU;:+-NPg'^CHXnSI1Wdo#BO£ړDAfsM1t5Sa1xq^>{09v4N"pZ)T2ag:~rn:s,Z#s9bz3F.N9)EEE[q#PYnU8n46fC{Cd@Bm[~s}a~@A(iP>G޵5q+鿢۩R4nվS[{rVV"QZr? 7C IeRPAF^}yp؃'/*5L W C[?e9 9 䇍ɘʮǕШ>(>h3bJ@ƚmPQ9,͚ YiT]&K˟P!r6 yN#k1K;mH`Jb ^%NJ'7>&ea%`2jdBl/OTw1.*n5"l8NL[жnkJZl /F4W)U>C_92r|~[ NΓ菣菣菣菫_ ?%0F:/ P^b\NcZٲ q|p;PVNK;@EҾz-S0r~$I9?$YI gWqT>,B_403?n5,Kp>>ܠ;7n˷/n~@]?P .2u~Ƨkr$J2wKzLUqZ[^7烊 _G}y՛*3M?4dayDT, W\&yijP7<]/N9Zb=S9Y2ZbqyeR%qP!pPF9m*^p%S\V\h&WZ mWPj=c21iK[5q$YЁիÏm`]qfKڍ(BP0a nڕ/NA_\ T5 Uf\A61X2g1DkTaPKբhIo<!Rvޚ FRz._v)TA{LNLC1eFCg.HHВ2Pߦg3~]!щ~a,c6XZɪIsGVazzn.*gwa[ċⶱr+Dlp!(˷{YP)|f!26]Wqy0 > RH6El9ML}qqqqUJ F-׽c1!7 %>`J Bhu:˩ e%& ܌B&OEz5r83DI#H0m)$qϣC!@B+9d߾jQsLc.M_M 9UڑC!PLz' V.qἐTY`XRKZY&c/>O5 ݎ1(hiN,/*@nl).I3D:ô$@[~raRE N̓Hf6_'a'6-.+9ΓҶ ^C2(OAFEEi^RJ-v낣 JK m 2k[9BR+BIaIe39ቼ#8TyU殖BbxFJ&s$ba7\ >L_(V%Q)j)}|-qv8t<SNjJ/OB|n{FuA!p3~!!p  =nɋ<ҡ{8E1v(a&Q US J9+u)}߇mTZ,:,>9{`$!ǜDL٪JUAw_<6J[% LQ!hgQ2Ax)#M}])n~}ICmo,At` ᜭ ׿MȘ>z|'o<^lm4r5}m(HzoQ!f߅khA$|umyGrȖ7 !;bZi&dNa~fM8^6b~1&xZvmLOg2t؎}>[ҵl۬|Du*Mcj+EF!Mdc.>t8cRʴb+hU2 ĪK%}QHknhmjܮL>4 І\nLtv o'Jۍ <|n# cp&1zh>hEyO4/ Nh+ƴbR܈VsR-*q%m="T3N@b^0#F&$n\͙P@jM)CC K` p|BQhOd@(>@8#BnVK#|C%}$%g68e5["F9|O/A(*JWR 8.)}fr'9)XF&x-SrBMXj4-<”(̲ ص>vukTZζ6qI ``{H=РezH8 d+b"5Q/hlwgTF:9WUfb(HAx33R R@|'犡&12CD[U@R&_CFRw]q \/nJ3ɯ,d\JSŴ{ڦ(*Xh-sPDiIBd zV9B)8:zI JGz)1]C/"rAFŤ^zw #EiJ k(jIPQeүkR^iF^HAZr \s1^gepXIĈ GTAߍrqjm`DJX-h`^UׇJZ,$EAdATۉunFokORQQQUi)_Pku()u)A(u0Ie , [@<-rj;|$6sb Q^;%HM)Yʝ[kBTRo- r'@eD̹v=J&ΞACLRwAMn^Em^US7{cW[:?l PNZjz#eFL+C9PR+w-ԸR=L)Bn`F~bnP#pv1H<"v\x}Wc@5Վ,$(TYɵwu(K0yG3@VhXԾN ^@WOБB0~u"g5\uֺd/>//pL>Ibz?-hǕ'K,xy|?Χ{G*ǃ7wKYM\ۇi9b1xsjp>\⇗Qlv&_n0pAƤB^srĭ10'Jur{u氘k6c\v:IFsV'$ ZDwy$~O $Y^vD5A }ã_-{d{ľ-|X}@nK' =ſO&?xzݎ~x?7dr[,'q)&-do"ASVN}`*kh-u6vAk>3iPٮޑk241V$73<-姇c(?~ DhhhDۼǁr@}fLJ~==|$@ר8:F(N# ]W:Imx;BSd$Iw>o8='>ɭ?ْ; wlMU*'UQj{ s9m9$~${.` 2/̧}B󑄁:~A`^I0Xq|kڽ{iWm;:o+)SٛIo+=vz/P% YrVu3rmhOݞŷ'P#VYGWߧ\,bbi 5II ខEr< %euAF_틑m="X 'vx3OЎWZFr!{q<< duTY\PřڰcBX4L[ގ1&xwvu[dT2TjwPOk7d|`do qTovX~t??CW;7: *&v'j'Jd8P\K?2}4Yku V@6>N>{/4|VFw殏!m]o!LÈOhh7Пr4\]Ys#7+zq8P-olga;@hKO}ER*^"uҲE֑DfH$|a:,B6RWvN hoY1 :[_ۥI] 8Ck4G!ɓp.5Z 'F ,l48@49Y0TbL.&ExU?ٷi?㧑Ƴ\;_..o=pAw% wc?-?2NRUNp,٭6v6v^Okv'bݜ64dֳk\bwg*skGC:L=onn\^sU0BT4o~B @Zr`C ^C95Kn n泟հr^`+8"-wb.F5xJ`qHdj4WK^0]%bA0 0Vo(D SA*){%P`Т Z8JqAA{ ,Ї9\]EC@Q 0oSK8ƨFS:0Tj`ƠBu`!4,x9/c)(I8-QO@Cv61yh>و­/|upk wϻQƘ Z~MY~Q`Z>ė_8?3W#WG_# lAOt xOzLg>zxOp}pC9-Dhy9ZnX 9FfV}J~3[D>(h(ZSvWh 6XЮX⭦k)#'@ ԌIp˔cq2sucO@kV܁hC[0}i h y ,_r'.`jt,ng Ylx&r7T`I>y3ř74lV>-$A9Gc?W/N2S.N=}.F+b1))T^;Kkݕ2o$b/٧!UeiѴuM*'%;M)5BNOM<# kX`((.u9Z9!1'#7۬BX#^C9,q<^7JiZKJ;g[ ,5$4kWN9FC r0,N51MRDD^d]t^"o{$rA 0niLGQ'ݙ3܍nM_z]R+'OIRѳYQk1$~2XIǮ712g~\4bRQ{MU.O7-f:~~Rw"RgҴ<³HdC>vaBh'>qu)ڴ6G/-ќRkl 6X|&B)B Vk۷X?ȴ~{n,L[D7BKfd"!) d@Qq^ka2 w! 3sj F6IBTiƈG@{ FD ˖5nn[VU=ڲuH%s%;r1Ѿ0$ kOcyw@kK8-qomCFBHd 0*&n/K4.ByxMCr)o؉owL|(( !kbUSsHLӍǁR³^Dhe)A|p$\wV|rESOD:~~km;`ဏqt)H{qʐs]7Z7iaOh{+r[Or,GC7.oHpCE=kincMcoj\Uq F!9XLZhk,Ct;5)fLu(s]Bw1><_ݻ.kVݧDXU'뼾ᣂv{gt-շ(]3A58LJ GN`o Լl5uNw %d;qc?rʼnF(N S~׀:ʼn8uSpx{ `0OїV:V.<,oEjzJqۚ5UVCB$A 2^B:,)SD m b=>r5Xă3Yﳱv.v ?2]4hBc0M$du5k5e&0 :A Ɉߔޱq8B $Peyɉ*h(5FN UlNɯ$秼R? atm+>g-swsݻ>biݞy4H$Vrw!A[n~:[(I  ֚?"Tx1c$vy~TqO!dLYl2V ">p!ݎw/4ΙNgEEʾCj*e}5;1t:(sGy#^Қ6i.xv91!nk‡B3σKGsqCN8/ ;土Ҍ,; (uy7笠CRٴs  :CL՜95yuX}fb.zyy(Ruyٌi'|cnL;y >$*f{$I,a$i&鼐3)T?;2r%}+&$+50"©_o_T.8W?ǫ ͫ~e?d:[P=D"OSgdov=(6"%(&ZA #mkCΐ^[+7!@t'1h{*c h Ɵ51p:IDmʕs iVh0]iA @zpn6NAQ0IKce4s$%ƅFyŠ̀lUg7".~Y[M-KMЗ^@glݢ57vt͌wg=kC.!|\P!EJ5$A$l 9i%E`S椀JR*)a0%fHSE*+SE:,f@A2E:)I*0x3BxG?'du$ !0}$fECLr"'4z*e^1cb]_j xF>蒾,)c#GT5弔tݧRLhuF+AjrxWAko JE½\9|\irM$48i'H&QrhecN\k)bJ$ymxG1S1?92-2،K/'\FUFPE" G0X']FR  (yabD dO'wYzuuP][Lgm]i ˓::v9NNː2ZO@p$ZX0Wkݺ O6 К#X(P!iJNh,2!YI”g܊Tc+3)Ry ğQ=W0"N^]ǎzd Pޔ9h*[gŁ#h48n@5dQY ✚Tf8IB`ޒ1i&F8)( ^zfYo{:6;нԓ:iM@#*56ÎX'-C$*mfd&)wβ^+e^k#p@|n 79El6,cePs#fgW\ eGO_n8I@Y- z'mc*W?mЅy2MG`1O.Si$B-4`n^s~1) Vvݦ&zChLn>hX BD'*p*Yg-zڭ E4FtFZM#jX BD'*^p*ݢݚ\Dk'|n(ZQQתtViLvU}s ,K W08]R --WtY o>%EfS{3t}oһ磍Ś ޼}#j3oM Ǫ7<BcHCئHJbG! t l+Ϥh5 %C+niS~fGo3ߙSZ9:?Ηg"F&ZMNGLTZN{~;Z3m{ÑXgCɝ?Ȓ5*RRv ޶_q *X/3)N:w7J\k5Ľ26V {?/O*!x7[U ^%E쒴jͅkƺ^:E+,h2욪mOФk |@u&ƻ?-4\6r eAn= :b=WFz}/@-ֱ4V(^th h˭o fo&y#}wQM˳?|2˦&oPBKdgklGotXsH[NFfH8: cx}ڥN1fMv 6I@uVy9jQxiw9Eq)Z,/x)b%y;[=$먰)yCXQpS)n4j)4]Wk)kC<4S}]upkE[ղ]7ob&E*b[beJATKu*V?\s{Mq"PĹӅ3YJ}HaCNPk6 觧*S6JTkw-{^o.vOA޹{G!-]g[лEz tն}4|^B~(/>)oAhOy#М TP$rz -^?Ѷm{ GXG;}D_Z?*mtVe"6F7~FPPi<蝖;ܧݍc8T`6.qlj{+%zIẍwƍDF/B'g(^*ƭcC>MzHrǘ k6MltӝS/1BiFm3ڔL}y Xk _Л K19#k9Bl , =>-+^=޿Od23lwIl&Z7Qk > \yw 瑌ȘDdr2vyf9Y|J_}ν͖PgfZ)gJZ[)9tҀS BrNK0j$g Ǘ8qg9xgox9Wro4GX_1""%rkIU9`2 2Gϋ$xjr2LknK&cJj]|fCIGL['93)0o DhŅ~2FuR<', I( %6FH7b,w:(QQ#=`omD4NGl.nsZ`Ð5YUpJw# »S( @Bg8^Np,[\Mi 'H{\&78fJznVl1xyjJs ) 4<99 Qk_st;iVAy&{齿~kQ.ÛO3gn <8ga?␾ z{{O~vIky9k4]l^UD6]'t9/zNe) Sfճ+BA+`nܶ> n6NϾ,BBzDe4$!F#_V`DAb":mTnG4SѝvhvkBB>s/S2=wz<5OP Oj*>~tHm\@xIVem/p:F:F*V|zq?\0Z HI aC un+'G4Z;S%7HcɽSX4R'Ib,U>˸IL63, &\Le r_ ֳֳ^>W_dN~4p4fӺy™ 5ِ!)jCZɸJPeGo֜})9Yf,Pl%ER椀JR*Y*ݪ}ĠN'm rW. v[ߝ>`R78> ϛIj70A~l ˨eWBbb [^XH#4<2q>aoP Zz|;`N(҆z{afwv4 {O";XS˰Oѷ =z FS0 .WMeΙ{n][+s3ف"HjxtKSk%pl]ۈ5!I.0 kLrH9jeRJˁ )I]8 cCc+!QJ0Q\ ь#&i FIͬhz2LJ! T #2)ܩ,X-(4$Rlbd[Qy38MI-Y*0șƁu'Ntd:U@ ^װv>iyfؽxǂREu y,K#|-_~w] dkr}M.=\Da?3xӛ3XfrE}]o0'/ߙ.17OJ1Z}{pw˦k0Ί:o ]T,bfjZ-gqb,y,AYSR`o (gXTք^+VPH_byaޯכ8!76ϺN1՝сMHg.Q2PﭽmH[,!Fv< 8U쾶[@C[hLq{~KIAU[,!FvuS3daNz)K']AM#9/j s9п ^z+R& 4;mDz%C^~P,4}ui%e'%"EQԨ.TY:$/]=h_~\MM?4;4LӼ3M47H6Ι < J*/T߾X#;1ZrN|#0 2!Qb`#UstJ4p©(9a8EX`Ė0h,N NfR 1'L2eNEiH gu*"X,S0 q $ pbS6aK#$WOl~"u3=YLڎlֿcFQJvQc Dh8YFq$"v4R1!vG=Lz?.Utpy𞼩Rrb3$ -NSepmm2ʬ .dök`^ [W !4~!{BMˁ` /jbބPR>z>U׮;!v+ >2a}sN'뷷4A*v *5@Fi0.KY5xWi3zub(OybUW7(L.u\t:]A.m;h(Oյk !| ygF6*^awĐ ;}o:bXҕtRjV^ػ%[0騾GjB;㝮jqTgmbn"j9H>E-g<9*іj9})z)90q>ŻHJC{7T)KG ΛW0Jq 6,Bg[z>ZwD=ڔ` @5%I\$i|NI J2 %2)vJhǎڻYjķPrT\TjyO&wE,s7:GbǗ.۪,߃\/S\*MFnWSҳe9JJ Zth9|L >^ Gw]Kz0#\|llYO|>yϿlE'/F-W; PPtzX;!)KĺңfI~%_1p@L'icb`t<}Z:VZPe\%B-$@ETE3^Ӯ[,\<LhDUzpt T.l0"Cd̵OMyoQLgh_?66跅s>FbWa7ϙ4Hړ"=9~tI& tZw뷟4*0п7 뭧 0pkof7' 4]C2i9 v_7rgLRzve_W1ir~i1eI}",q1+4 R:rDD)"СnϐTwc6Ǻ-.n:]ܹWuR(G>'c0Gٰ&]f]Wi3zXpda/ ՋU+ ItDܽ҅#X` !|"L Gf2,!T_Mym86Ys=N=—As&Dfh9FG>iWi3zsd8̔wҗ/Oa8pyq72Iibunqhso$(8q.(4Pz@_ouIt{VTߧA_ f :/oAwK7KqwWz9I&|0U2q;n1FQe $,(S\.ig9F48JE(uT,Oh 1.~T9aE秫so^^4yRL-2[D"ϲ$"aj=:5ܬãokV#,_^,1aB/VMHS2$f*X!Mi)9H0c(I: 89@_FZ&r I(L\Dz!9TS=OhTe"tDEy(8%8UZ@yJ'6꿾\Z_&cS)kw,u\5RřMUr/B`,,x⦊鏔UO,ۍZ*}R0MthQ,^>)YhLjpƩ2#  KE"IQ1=F8c*zE9QF$`=yWi$Bq=N[CKoz]>]<8pqZwTl)w/lj8NbT8< `D40#fMbO"Am+N?Lh N&7IvSt$m}RdCwx8Bb$80#Mt;a& CNwʞ"9iF]xHD[<$Z]ܔZ\+ 9nJ.&F9_;~()柳7Š[$31jRer8MjŹ I-FMη҅#X` @C Gf p.ՃSzڌ*0qyu+}-M_6Ys=N=,"\4/O_EBjա_/M_jk%3f}wଞ6C.2^2FʖO{͖>p/2[o7'Z'sPuKy&j$lPPb^Tu(8AfYgAqjK 40h~XHpbѰ\Aq¼ɹ ŰQA!!}A2ZSPxhONf``hktH->{j!VVP  #ʪvy{o_UVjfxSbޘ52_zFms~^{J?1Ug釼37Oٶ(aTpZr&(aj׳{wIyqSOƋƾ웒`g}맅|MԫMq6y\nT>sD拁M}wsUڻXz>ZwD;)h9K}3%y\>c +]Qho+]8Rf[ɂ)ds= *lrl Q"^Ȑ:p)LB૨B*UGzoʫ8-=-uX`44 8/?r 1>yz=U f٘׬_;ݏ7#(O3 E˷aݗ^gNՇch}=F|:-P^:AK<FGg^QB_iڛsҖ7_|FOFGSy ߝ 0 ;հUfj˟>Xvxz6ϿbiMVnۯC"DLӌh'^!k?6$!x d/t CV,d:|mjNj2#<_͂f_n? \ƒÄ.QDEd@deQJ,ʷAȋ͍x{ր'PFY3*μ+VCPAf`1i̛[m PŃLk}A?5%-g?}L efނ<". BG9[>O~ 7o˳M$h `^WZ{^^=<`"9g7Dʼ[ <_@mҮͳQg;N>;68\6^ AwIsy.o+y1"yz}}ayqX7CPg\j,ˌTiC#A L$k& 2#MFRP9T ߉IOSjӃ OSoƹ L ]NM~~?$wFKD൥ν,qaArl:9#G\0O.!F)!$V5H8Pq TעqPBXOᄰ( D gEQ,!4pi)2,8i{J ?E@98E :aDdRɘ8M#X$Ie$0bQ Pm1fR:pKN=Yj>2wAſJIBlAE@'aGrB/#G6"L䴡~OeϏIq#b"@^ay5 -8B 9i/?2fQ ibF8ʲ `JFZsewչsv뜳;MA#DCXznJDGznnↁD'/6S`l?$v<>}tKIb#Rn(l}oK9KHh1X $FGVu7YXB$\(2 INV&852HҘL5NN.G I`d;4 ::`PxsVNh5.?s}x,eB1tr`A6'ĆJ'F I coK5bKMo~ #Р`< 3P$}dG1 "Qߌ6 L VMp˰~ Я9(4Y$.oA9> Tʹ s.n-75n:{5>WԹXKw\\Q%e[,eC'8t38; @tZyg[<%зYÖ2ޙ-j &3п8gzǗLB{9r…8pWO"%c Om=o%p)&CTEr(҄}PX[y:Vʉچ `bEQ&<&8S*cJ,N-JS#r$5~R=V FnTE3nWYwyڋ.~Q<ATԸ0E]Jv >:Uk Fnחn%$ת&l>.B6rwh1;0g2?cij|.s0_ d ^{H-P/6l?WK`BId߱t“Sx"APH&;9 sf^'['r1.u[-M%15B.MP}cJdA%!vxFuA fBhEMD}Z>ixXJ"dTY&kf ,ŭgXKPQdL?Q=jRjhVp Qf֊ E{V+8"%jfN(3H #C23+>Ӱ%ϑyCJA9H=QVr>kHU[4jz3w!N1a3p|*oͻprT5m PNhk6JN1&)pDƋI⠗x, 1y]yF_V"7IK^:}K~T.-r&E^AkN|ruK`8E:ê-Z6])K1"DR 2rd(IZ'd52ST0E+cOhW:ULK܄1Q!5u׃w{p C0'r܁VRA(jHq%!ӥH;|./m:td 1Ue/L< f9_ڜЬG22WY+#J3Y:`!1ǿOt|)XRin SxѤABn?\E4CGX7Rjw 8MhwGT(Cɵ{wDUrʜ7sʽPmAZ\SiޣYV{QV%)(Cgpg.sWdBgyW;)˔\=K4Gr.$;=QXԁz+D[ZB9@Co#I_8U=je݅>su>遾;*MB% P\:]ߓ_ J dW=DdVG2ѯV_6!}+.H!4U,3& c#jlHRʼn1&RJ,V:9Xa@>LLU/v\A6M^h&<",63ߟVȚv$~-Taz7x0eŗ?=Hr&,}|i_~2yIAWh!:r7C0Ֆ`Nf_)?}f-WN4̷@͠H6`A2N(/?`a=Rb4'w`,xBQj|cq)aB:q$@Og͇LrVl0|ͤ\R> a#t=_]›g*6Geщ؈<8u]v}}+Q+7{{ƈ6>*k8WA:񻲴; [Z X1բq1Q}N$R4 zۻng rJBRUvuB#QU4,] $Mb5kiiT-ʊ+Mn:nyKmLun,Ge UB)-aʙ&F8q֍3Z,%FwT̨nFݽ<{48T1+t/o x%ύL}$tO٭~vw@ !I_U}in;U ,y]y2dw{֛KR7K1#u1✀; [r w0tW=k7pj4F81VQss5I']P"W2`kXb8\W=4o3_O `ec*[0;6 C˓yc\u}up'Sb=W{ CylB=Sm15 ݢK dӇ$x, ENf~zfcJ3?V IUf> @b>/4fr'+m\N^٪OY[dls#1WT ?_|w0'&vGF%r+q&fkyQ{.Kumsn=׍7mtt*aʻ^6Q m~h(F"|gnCJjHhL4a0^xb}>b>Gژ_&f`l#XБI 4Z3dygY WUMD?%^MOUBrN%SB/n|=R#Us~Ջ۝'"DVv9BBd^.>9JN#:rS\կ0J#h% u%^nn9Tb\j(4!(Aju'GJ,\*}*8}% piޡk8ײA Yt1 Py1" "N7յR<>{}G}К%P 2GhjH zD\D"/ $q:CUz?\GQr PZ3x3FͥeV-{i=}}ꂸb+D-ML4FE[@Bኩ7ewV86*̯g 4lJ`x}n;m8K&涟ۯ_; 2{ S{ʨH0LpIavtU&:ONT1YN bR#ӈ 2%f,1T2JIth$ I~e)I(tibPɘC"Piƀ,ePdѱA)(1'txՓr^h  )t7sXr!1Ғ;ǡVYrI4cZUɌrRMb=A!ӌ 816  S'}N0rT9[|z8ISX(S>1;]\ LR!"&SEE\SԒODLuš*aOi'Q:&u-13ߟVI+dgOjkn[9*&̱㧷y.x>H>~x|4XxH~Ẋ^^XFD!bڲy=?S<34Z3-P?g3ho,-gAj!1/?`P1L a)p{2ϟ_cK$Hc:]svcf_z|dd-hr+xňh.(T̞Lb!kknFE=]ɸ_\Imq}I hD\d2I?) n6߇n4nd/Yy,z#y' NQ2_V9BBo!ʱlBzA5"IHl>3|M&jv;pz0or/R/.(:a` p|,F+vG!4ZT~ AHUk4nE"F pY2,n08x I0 )@!Epw?OWۼmo׎l("ڟX1xoI7mO嗻S{LIaFIQi0;_U6WzL̯0<.hW-SdM6?,ZMWaqliݎRL8\*vƖbHVVU1Fx[G [Z%fT@bb^"0Q3 j" f kܰRalYF s^ITFJEwif *E %8%L#--ڇR NE,=`wS;D~*AM 8z BS /VlW\[yK(H(6sZӯiu'*A*~:w*Q=\"u;/,m;ub)Xg *_;LGV=ZyޛGmӷFr/{~44YݍwGFA[?E~ҁ9sfY9JSӶh+͙,eǪ{O7Na1^sWFdҊ4nz%JUR?LR+e}ݐ;# tU"82BCfws奎&0oܦԬلNex,vB$Σs5{1ȗ)$JE1L$H'z)ñ0G4Zg^|U:ϕvl+6ZHGF3K TzCဗkxhelU=7>) J}7( i\oPݠbڠ3*L&hRe2[Xq1TQV:G2r-CazdY~u5]aLK 0'J?|}MY@N< s JMi$p 3U4Bm~árzP2y X:s/ |hgh~+m(|4h$18Nze̅Heb3ST,e~%%e!nF:j7C-&NyG-S4)ЦKJS^'QΟΓOϋ^&Lh.2Ɠ\ fjy7K / 8WdyمRdLOȩ^7ΞZ enlүAfaꝊeyYY.Ϟ3G {g")ejH7^'"N#!f"*O\ka?jX3s5\FT ZpJxG WZkAV{[im++\UJҞ A8!ciW^JT.;|x-  pgt}l.B|[]*; :au=: wZ =X~j}d^(Y=|@i/$mF sv΢]TH e予>ur´Б@3nSI <[A1<]߅ڞ~RxT[G{q$hc】t@k4 W-g0Z(`? dHhp:"P;_x{xtٍɞBoٓ Op;A/ݘ?t 5n;^c?Ԏun-pe+]HXT<8t!X$d+Jau'.Yďխ@鋎/ SܖPXioH6z6̰<5z  }.K'6O[<$@)%kYm4q"جfꗿһ_@Lc)S"depQ D)0S81,*{w*& zytN[|7QL͖U".LeBG@SaGf\X#F t /* :]x{Z ~6u><ZCzG\ @p&_KM|bfoY8\Ujꕡ@Ow~Q.^zxFO+^zk޹N<yh|%gTQh +rXHT-]/N.lR)u4Rh(oZ PL늷|6ZSrg:we(sJR;`Pa&x޲][Qpɽr2xgv% -4+'GE3 f e?;?;.}n2M~sZ#G9N"cZ}uNue ׶A;3r"@G?8j"zsc.8C૥0 %0HT鐠LXdWKOZdRͬaPLoOo4FEs)Oxڊ#x[MhZgTжY q';y$5nEܩEo8LָʩzWDlW:^3魐 :C/*|# R []̪]E; m 5xo+~7BCiN ]P7Bgio&sJ)A>.^oYbcI>Z#Cڻ>PQSmCk3G%$!t5$ZI u8A0$fӠa0iAT`>v|!N֠dy#О_pJ0GF}#o#!"ۜ\pȎۛ4`@M-nbLnAXno a?qaD0IA0^ihj$ > 7BB;z _ZBʬd[l/1OI]f^(xS+3bR#XĢ `0r*TD'IHBceHUfoH(cp*KiMO,XcF(2Lef05%TjDwJa$#ƶvhm'17A6Bc w+.pB^\ڮ[#+(`r+HDT+ebdq8 K!"#J2(H%$@rlW㭩 2M.zRFK?8,z ]W}hܼ.P6^뫧o6_/7W?Q?<,-xw|\I;_IX~6G?3[Cy5獀d3[J,!|R!KL?' %XKQő^Qި`~Q%1۞\4|zƭqNqi-2^kON׌] Sqفeۋ]2gζ|F]|ȟ.#Nѓ6 `/&Ztyv[4?r om=/yFuB5YƧX^K۷M7zPݕuB'P;VPΗpXqgo%(Nf*w!T|\Z֯a/zR;okK-`%X_J-d+-F~jQh\{f H/Fm˳cb0E*UgF!Z$l5I!cR3?]:* q\tGz \,;?rô]]r;ZTj~2W|W1drO_[Hx)Y}ԲD墘~~f_|:)A:ow["抁q}G:9Ŵ1qn>f_|֟SĄ9_O/6Wݼ|џnW5ե{v]TۧfW#BD#䍖`sC(%hP5$X-$X?u@bA9{sk03MTd}gAcZF 1|RUݗO0 #%=ʕ+T]q)nT}ZsO \YFdϢSST:57/z(utӏ7j9AMȿOu2ww7y%'FJ.i-IZkLfɇ ላ&#Q@X4WD%kw3hZǕ8Y6Gۇ(0M\d/q{M4$Ik#p!u\i:CɣD?!rPpsyRb8@xHЃ?8j:g#.=J˗f 1ِČ #&1ֻ8շyYG,k%Ě!-o\ATÂ,]$;7F g}Tx6M}^І~5&)Vth7m2Y*32sZ9@YN,[5NZGۯ8ysտOjEY.,tdr9n7H5nv5ު9T d}hXڥK08]}d6]zOR}:' >䍣d"a{'ު+M :Y+zӇjI(3ҐUg]k[]f|7fT>W"}~}qbD <3qfOvל~GSSwH@MF$aY/eDS~OY 1At߱82dDGft4q$4!3~kPk$s&=r:o'>. qk.,:˳>5)BhoEB3)`$)K2J8*V0PXpe9:H4\" iB fDRƠ2Kx/2Q"@4ULf s Gز 9yFWw(Uїh/Hf"ѿS{} ґD.`^ 6rϾ˿ذv9v%pmH*o'o<٫{ip{^1KTSeqAw%oIi^$ GnFz虻^GO4SKպ0q@mY4kg1X,kn883LL7Lio`"M%:5^'׬q~l 72wMN>D]Á{Q%@yoa o±grnfý1#|D3M 1ċL)gWrGNn*Zgگǝ 0ȌTfr6؍Ti}ּSfgi#c~(z,J&s+qxj2"\uKތ^#;2?}.Ny@H*D#aC}Q$G!\~&4 ]\ޘA0 Ibiv;PG&BaDQe4J(f"LK_2  3Dg'sr hQ<))}q7Jo C 0ԉR#tvb5j%Tehe.J0&#} d|9vbRI]D_ւ&d$lAqwWb3-sA/U^qmw]e֭@·I&Dޏp. \i;O &f??~-MuPI5dI.ܸ(xN!l'E8ۓ ƹ[d;֩/v8i~+}ٗ^0Ņ[2:a&IUDVr2%&;&yڎI[߂(R[lUMS/mGOQ3*]\X6xRKo\xNT*(#~&d2Yf&6PmwVkɶn}= JXbf'n8)VP ] &vR7GAU%RPCN8$D!WrPF0Ԍ*DĦ1S`?~"j" F pd*zG,#nP+˓@f|.PB80ͤ}UCh)Bܐpe/@ZO7Z؟eI+*1l/X!{DM+Gz-%%b륫,⻰fkIJ1,:@<%Ow|n\0<4bmG n溪, =tTK^^k(giƵN6U|}E 5gu#KFIE4mua˅B4k7z):~xܞhk#D <ߌR/[*{7KHfTꪳJCI,՚f9JH(ZfVJ_kPп~.ArfF-.5Q͙ [~2Z 53}:|n۪n۪n۪n;$B D5>G$D0 Tin7"iC=x:&}\9q4f$=)HlCZv l!{:Lqz6k魙4K7>`ܠI,^TgוkEvߗs~iiC@W:3$h8 T]v }*KX*xlL#?wS }6U w{:+%t)\7r,e9zwuY7}˿.::O5p4ϗGmK&XOZ?>;s5z-M7/[gIћV5G}> ,? >M:X2ԃ^b6,;HWӓ7kOԝ}qyT 9$7LotEZw/kro`wF^߻qcMdC~i!A̤ .Os#|jF9AU*@gő N$wq׵dćA:J6ZWCZ6}0SG܃cq'NcЯB7ߎ uğ1a{,>-j=?b`!mƗ^K?:i~:~Nw?i5N$a'A&`:=(5%w$U )|~F杁:~I,epi䩅em.R^A?rM3u7sw::0[+,^sn6Qf o4éϙ%]ErW۳6(9.g̢,qPf63ٍ&S{Nj|s{cxၛ1b󱂹ǽG ϡ. .1LzO <"XEQ>~Y}oˢCc48B1xQH1&\P1)x`"LDJJt6z>x "ςW)X3P**Z kRO\Sb-Rymh͸$@fdxR8Bm( I³b*LB> 4 DJFb]ZR0̛+ɐEW,A|O3(x>Pތi0P,tR*Uӌ) ).+R\ml!5BD [5tGc1 6#XC'_ԮL9 Al6=t#\ɮݥu-5/Sd:a9$x]N0B&NOB*Dsئ4c/Z͌Kp:Wx0'},zBb@bReF #'W0RoQ"Lh8yJQ7Ot/Q~~Oƺ1mopuPߔ4NٻFr#W:чݱ<1c/ePnͰ%YzHQ*, *ݒ(<d~ 5ㅮ+Qdhwx[-oHiŔzÅZ.r?EW zY!Q:E׌:4{-#Ys!l5#Ne{_/!x`3{r5JH3eǤ/6JjͲ2^@"16\ e/FU.;oEwpf!1PGivig&j14eM'pSll EȻ+c(GcAH9$ApMѶZRRx.Bpb⡲ '[J?7+"=#YZ{c[J&|.u B ׶y =K~ɢ>WC[ 8 h3XQGgJ]AXHHwn()kԚKe]k"qﰬ Xe%5)~qjN *6JZx[agQbeLji .-JLQUJsjI."DIȋ@ް&CBY34Ee{EYr޳\C`I9B Ű#Wᗫzf5k/Ww,x#>5Zg' cH cBFO2N>,Z"l6s)C#Y?9)Q4>Y(Qc{`CmϳoONR_/_úᓅ4\VGqP 896W)땸gU! ~^nI{u{&7uJ`YuHn*qmaOͷVK_EZ`]ӠKӿN|2BvHxxkExt>yrHw!SDG?6<؉NU ڧ2|/rʨMJwM.VSc⒭H\LMKd|H7:Ma TEjTe}H\Iu;VK]h30GS( Ef`jOFLv.&Y~|_o+|oO'?>sﬢ5ӥk?JS+fuEj+*VE Mi~[A';{27sr[S!6}rIH|dp»}k7*1`^EK=ϹV"IQ\#ED*\0[EXɅLq@JQʕ?DXӡ(9:y$L|J=:3y3<08&#i'1ëb1؇ ˹kA,YX$gaIΚK=$:emU!'Jj^Y?`%6,Ȗ ,)yV};=ܷ7/lEmeLcảqB(|l ߾oSzH xҳ#_*-7ɪBZ[hhxҢJ"sHTbag#ןoí㧛 (>7LAk.ȼ/OsɤN 2pus.]|F(K 3Q^-}<>E6 Ԙ>HHi>x)EH|qꍓU3gSdl)pѨ.|z!I3l˵wz-SrI3/ögSFUͪ 0|8 \;*+SP\[Mh) p}8`p D?= s64oB37? FbHXW@ED R׺bRՂiE*.P\ ˊfv")N9%LEVˇU{(9 e=8g)[#Z+kueGkm.t&bx`msBqAqZ>VnCZ>t;_+OA %FƒG[T8:ȋ>i~ -x%-$F G{jJ4|_RWkQ;?D*ÓF̰/ ,Y UD,X(KMa*]i/pi}Q40& ozm '{Q?I88㦉bUlU&jLzPKoO'iΕQ=I,$]1 PZ8]EoK,HޑeO)9".wdcٳdJnZ{Z!5^Od8WJQUUc%U)x?XH-o>{k2EMYB1%]]PH ɞ#.b!!};@tiHsD8 N4"dm ("=\ uH ȡȝP$$׸S孏4D9+%F J]f߮XD)V_]_->ur(`zj Joj_H/vJUJJY^¬1RB:41X$BQ8E4!~u Y#r*@=.lZOIj0icf7x FNnS gZ@qm-L Y|o2ϯװ>=٨Ϗ}iJep$^Q?l>vWa `͝[Wf>~85d|b_so$zx`n?>xU-oxy{*AǥwA?__¼`e4<)~1uF>ѵp7%Rs8fк"(84F̂z2 eT/MfwZ1+\m?VZkԮ֬T Pi*fvQ/f?[~I (یs>ʄ8Jo眏>_A Qu@EsichK$\jO8!S.˺2ODOL Z6鼶†j&em+9PWKHVR@wc H/20n>zcPզhiA(3jRD[VgJxO x{[H/ұ=o+Z؛? ?z[RXB,ܱܻLYN&bcDN%-j{1絝?Pdog77 U؟m%迺ۤ}6}f 0cx'<HW!J(RU[Z[DWX-}<(f.bQf-F8c1GuW#n =:T l}:6e[Rm9~yLoD&n H`5EF#dj,LHi63@dk)9z)i4Nl"<+D;HeY`ZM3,EAD(9QK}snM]V~CŌVK~ ?\bR/2> zKf<(|c lTšy^PU0@}Jc,{ rzDyvD/٩9:)#df4LYbwvW^CrEd=_u߾fu_yqpocͦHx@7SoS%K ދv-Png[x=Բ=;l wBMPGR]UhWLCbw!tNSx=+MbwGz`"\P:[r<ӼAv[ ? j=Br%9`G7'|pǃ Tk!gf Ti%}F314f.(BW/կ;y]fJ(17vv=& yv61y-/L H2y^n.ջLo}E4\A@QZ3Zt]J~C FK~:(#Sg' v[i?V(t[x #w:nܾ}Zʖԁ֔cʻf^UըeiB[GΘmkSӆl,grʼ\%yxB,o@J-hqs3\ HٽHRtQebY\l?tŎ7^[;\to1")=Hv{'νʛHʩ@ӣ}ULH_㢽܁Zsy*q.B4@c ?~q oec䡍SP#L|1|Ȼ|?5=Ĕ&tX%+D7HZ֢h?j2 sP FK?Ɣj{e~jԂ!i3{A!t )/Lۓn=<2Pqmw/fi?=]R B<x=~YB?)כ2Y(C2Y(C5ZwW7ExE0榒(1JXgJyJ:4%\몪fj6{Ϗ+MrA}6+<YAoV^~9$`Xr.S|ilh[f*օ]\z;`c; @@v)A=e"5bp;?|I>Pi,MFy^OqقSTkVtw)N(6Ͻ&y44МgCs y,h1(}& C\90:Z]*$Mt||Nn,kpsA%#+ j`u ~8і%S"qCD x^k] ]3B=/~FG6w!Glj(FrRͦHZFgPTCd#civ&P VV+#:I8~;R!VAh^(~.Eǝi半=$T NI Vټ2\糘<]Χw>`rE2Ceې㒇O[ ,xO 8*ibIpȱ䀛䱻*¡N)`lELU6<#Mkk(Bnh'g;K8S( ɑ GgvΑ Fòl"zdUB_x5$ X&JU5h (X49b1a :\JQa<!zEJJ +FQ! _a;K`oư="]H_fm#(7(F$F 2!xc,6`>kXqiа5AH> *A$(;5,)Ua#bϣJўEbb܂Q䭶ɖloIϬƮ^)j'gse$.Fi1ݴJ~c{D,^o .OV'!d-;U]'_Aцg|nvȍ*ZOޫ`6)oidoJS &G|% u}v;z3XVХfԠUrb`/ Ľ5>Dӽ:?qh}@NN+8w'Z~Yp 1N5cAs{ /8cfppB_o-YLD 4%>3o>pqyG'vF'ϙ?5f!𙸞\\'%{;%3zSr]OogҔb:}Tv,K%EWӋo_ vk> Ow8mtSYӞ"kkr#"dcl >gC ܥvH^~'djɬWΆkW~ 4a=>pfkqN*utYiJ҆/-zYg~uyS7K{TvuJ$LLsb~m*W,E06TDVҢTZKQ8 `,93 磮1F[YUpAW.6:U?Ϊ`wo6缀Ʀ^ h#AӦbcO[+ IJ2Թ)L]HN`WZλ#σEF#"Ƅ[rD@*TuJhbuB8ul &N=u)qp.jT;FJ946JB+-YRh^p(M6 YtcX emDr":2xQCdsw$kjQ_K79.pWV}1ֵJX%'1Xﲀ:%<3A02(61udÉ8̐:xSGu 4P1YN+k\%oƾv/``K _}l"]?_qw3f&;(:f_We$8(d WbI5ۚ5JJ,ĢM1k9Hȭy).cd+@PiťX zj,A!JvX EBAJuL2Z`uoW°\l˥mL8B,q+D%ȢְF,ASډ%FZ|XIwuMm/ZC5eqN.0SJ ؠ/q5"b`*`EEY38C% D-:̻vN6VGCV.](sEp(m\7rzˢurBtΆ$#$08yimslX~#X:3tµ1Z9پFsVmm޵S\j%z 8E =D :{Ts{k>t@V#u`;[<4Ļv˹te;'Y3CdNg?t@l6ͱ:Lch,bK5't Y%ȗoNP"z+=KcV _q  KajkD+ 'lk "B!|8<kއ G"PY+E.OΟ.Hd3\?YEf1S8yVl#>Ǎ?82̢3 woi.P$ܦyMuֵ?O7'w-F$cR9: tw;9h(t6`-uTt;Ic :PiBOR [BFtDݣyT N׵C툥Tdg@5Ȣˑ}wۍ$PnOŮKہR[doO5J#nsqߝqTatdհbcZҟLTpj/Sm0^||ռϝf]ccԂ,ݝ nüQ>1 }3{v1M;_Cgsg r_hϘS76/od-V)-XTDy̚jD0TsWQgnXWMw}.=Si`uG2ZOʔ&eӋi~r:$(-Rl3/I'ڐsrï s/x!Ap~|ctws+;hnj.E/դ|!sFW='=2g? ~=;A#K7dw"7Qv# }'Zɽߖ1'C+;WJ@ 9td5`mѿ\+#P|p{ޘR<4WS>w; y۞)HDJgl3)M~2x-C7njG#3{f;?-=RҎ̈|ԁx6N{{NNu̍BqtqD?J;7НWo\t؈t z+'k0Jc{TJ8@qQȧX =HXNZ'mSoRoo{f?jE7%κI֛EcfycBv1rzi7!,ٞEH9S'vuyލTąS`5 _(d|ELB˩>)臭V} (:]y7y4)L;uwxBR6S5bda=?cy,4BQ,6@A5 Q{i<νUebdEP֠0VE\i_̑5D^3T^ZT,Y aB O2wb_L^ZWN4vVFƀ`4 rBnWš(76(5dY_ AJ6£FjGۭ PRsZDcza쒌R8w0cv;=մPXhr*plՅuY3WH)ɐ( ɑ G5`v즒QvGg Ehgͱ p&Lcҩ(S.-'4t&|䑍Zƌ+'m#9,C}ЗJ6GO%rIA^rpװ%uWMW7$J+8x\Ga 1Jdl!ĬMsNq"~0Ӱ5 ʟ?`js}ٌ={5*"PjT|;mE-0$ycZ啎''SjI~_cdC}xFa4Ub&yhS3|7oՑҾ,Nz.mG=1t#3KWX6M A-ƭQXs-tRv9Ghmb/ٚpgd'hvQS}l@GM=R _cw~, 2yo:/[ /K'XYk- AH,wwEɢ?W2 < p-t:( r&ɿWd a#H Nʙ +#\DUcu@2K,em_! ?p1Wg={}Zؕt1ػpL2_]%s% ؓHƎ?C|(\9y 3\PuG`b삅QI32ͤHE Ĩ1AӜZkgqP;ꣳ^D 'G " 4k~vPv3b6MAG#EAkəw6 8Y$:G Rbl&P+̖iM͗ݗddR iUg٧[iz])3÷K^%i &.޿!7+K{t< A"" ӳ_`zd&}5MS~oƓ f>[O@aDY|p901 `@ݽ_tǍr$X&MÈbG8`WB謯=0]RjȎ1YZS:/Sb6D9RSiFΣ,ݝ=)[G8 Mo+=VvcwWT y>/6he`wWrNgB;b?ܰ7?Ɠ\ҕg,mnzfyVs0tN-蝓ǡӌЄ_Ao fX6]TLw{1xIfI-Bb,BoCETyD^yͫbAXs9M4t}n?ɧ,lRt̕phaT%?Guy ZaBS~RM1{̂޾2ׂthˬZ30Ǧ2Gb F1^xS^=& .`yNbskVrxRSATs)G /4ډF,u8.ٍ6ۑ{p/xDd9wzwKt7It7It7It*KџvcA!1 +l,w F)%im"9A!At̏n,k LƩB4V[n H&IIq>xGJyޢXc]x}49KiPAEMcvS `Mƒe&j2̄l,3$p&Vj HPıR B|tG 8=y)sR"# ?QJ?gn)k?sqFS2Q8/<A*;4 lm=!;4bd\I5(t=:KXDEEzZu4X00\9p| :mksxACy :>%pçn1#h! _pG镥'p4:M>5ۮu'a.a_~ 'ʍtw0R=lQ_ #]~@Q\ xrSW oe8%؍zIhG b=OtWW@i@mq.VuE vvd5:CH-Ylߛ|;WN)GR)wcxq Ķx0.0!B0 ]>OYoW}Du/VZ[Zb֙k4ڴyrIP8 RqzLR-}Siu Pba#Q?%"Ӗ=΂XHӚ(ڰyMثŭ}a҄c} \|&t.Ǯ>5$:RŋG5懷11'%ZsGրJoz\3.v!m${nԧnyɮH )9 xC0i_T&ɱF6UT/<%R"^>@٦<]?Ujy'w#-? Kc,PІ-{̅ 9KT SZv =9e.7U(O|;X8e5T X;zs&&&V;Nr+m8bµEK,#VM7j+j RW.%Ͱ\ 55U"o,%Abe5-s;l'̳=FQ!f#P%RpVqS"ȲqQ -F2b,*(*k QN\elY LئC8nTJ+ QƨCzCD9(jUPrMf-ni}MsPL1(5 [;IS7Doۥ;v)n|;2㻤yjS`b?ȉ#^ w>}YWRѤ=rM)I_i#CzewRܒɯumnpa>[A\KN>DQRM- 8=x8[%Du2.87H)}:"+&(1)(`f)'b !`=&f HDbNIdR9!%ZJj ,#D5fx8=#AS &)[PX R׻Ȱ;Q ߿+i+B2đʰtz:[`-Z;}ޓwG՜;O:穹Yvn$!D%ϷӤ$HͼjkhFL›`a㴵xwxwsT*+YoMV$}EBGz=k 0d[LJa{w}!`$o?nQaBKݨQ3kޅbjS֐AyrV `ur[ضsvnp!p8 "L ;1N mJe[ DxewD"Ȝq$3 Z+Nlz Uyu%-&H;ŰRg !rAFPc RQ3 4} &6jvߩVXcKX4os[-K >,~w~q<~ QkODJe?%(q6AP1%ɝȹQoP#9Lj[ n\Tc\_0YGe2Wae^fjYu2ߢݯ9J\*x搌j$pͥ!'F:Wg?檃󻗧6Y>ݲ7Զ>7.'EBxǻ#I+~] p>`m6Fޟn!JZ9jZ!==/ZJ8ș﫮~Tu/5?CA9x,O޸X$Gnw`j0DP;,DeHB-J;" +v_%pASFIUoZpl`S_HU(' 3<2S11Xe& &A>R)r\+6W0a+\ F1-bߌͳ{npS)ִnpX3šF,#"bb7/GlG'A$ өH{mI1䈡z}iLz'~w~d=w XpN^3W@Nd Gp%4c#z[k#zEw`AC _5;q#a09-2>̱nF:]vIyՒ`sy9N"ĴOaA -#M`qu%y]\w[4s}` Sd?R^`͌@ܖUJ+Rf05RiӣRU+EUݯ PuIRJK!2ҿB)̜ϘhZX*bqY~$M}Ņ_~.l3"P1LjZ)up)#T#9%a p#%;3>/sP`w dW2y[mn(Seō2מ0̄8>;", [kPCc>Fs`˶`NsJ.d^*i%!VabV0]akdUHA !)W,,KxUݢH%쁛LPȘz!1,KTL+2T5yy1HS&sKa. M23Z֞gyVLֶH{0S&S4K/Z ZxF*3Wa5s-AaS *5L|GkWB ]/| cendRda[Iܞm19JٿŮ\-G̮UaIb00gO+@ȵgɳOW"~nZH[S ׶^\;`f?~dW|rs 1h F)x–F6zC蒜vƒAߎq^t$O9k rXk8B tEzU`Bz5W@' +`GAGluOn~*f#@QZR5W; n|N)Ez t.WI)W*#!RTT#ddgSJ%E 1Y-9_4e͚_ց3{3BG p+(*JZ2bBbxpCLe.I%}T&aȵ?-n,w~/_A-wroٟ \W_݆?ܺF w51 ΈVc LP(G W,s &y%*eq]DA@>vw eU j.$kvJ]Pb),g*7ђb"E ޡ`6֭ p )¾J;š rn[ `B׏ Dz ihh ;&:sa a 64*!%'yU֊@J*/ouiӢ+caT _"oZ3_:0{t? kƋ0rcσ(UY,ISf$# 6F0JVٞw<; yϋGB}w D#2T r" L E3#Sza|K :DŽBB~c끏}~J)dlW\q?sD/Eq˜;U9MD ޤdVɻU _ *mR(,c,rfA$01(+]Zb*GA*9~ʞEaY~2$a(Qbs5A\(4*Rd͙DRڸXD c]n$ہ1=qڝ5ղ邤*UsNsKtY<Q ĂVxτDZZ~ujQ`DB>kzԣ17S*dNJRZHNJ WUZ幔[*y v9b`))̕tS0z &NN:CÔO0`ゔ[XP9-I`!asѾ$ ȳ! lWFHwWcNҳ$|OɷIMo(D\SZ x} 9w\'~D-r6rMw'Zmݿ F|l#:VVq8_25Ql`)>"h?;UsQa?_fL)W}Q)i^L "1/NJ+c.5- Iqb0m!fӰ()AbIcDm0kh {rKQ"sU[. gI3* ֶx._4dem-ݑXW4ң>z`y_{:pPJ[=شpsށyH 2}Mk.۟<1P{ϖ0fqݞ& lxbEk[{U{&D!xljE ヌ)0+_[67h''"[nս5>m6Smi}x_3ͧW˝3R~i2Ylt(+>m2Y 0":e;%0d S-NHwK!JMgMM:S!CNwa)lz񩨨Z2f|NPj/DB~%/GgW:ǿO@^X6vV^jMs/or#X-Wu®a]n|= {3}PύbW ZGM4ɦ$L]n!D[*!6q`-=>лua!DSl c1wSRMbb:n"S[] n]X+7њMil[,nMpzfw\(25O,N$k}Ȩ@(m{Hom- l\,飃w>'w^ {^I(o/LLrYY'ϒ3*oW1>^?UhLϟ`~˗/7HH*RBicQROtMQ^JҗPҪ郉=i%^#ep%hr2tἈ.cQҥxJt^~ƱnlI%8zZyo>3fY=I{{SImٺwi^kS.v"5y%i՜LxMemW\m c.ߋUxJׄN3Vy9+32;ٞw$5ʐ9]ܶRU9~ل0r#b٤8n?=+7k·ָ|ˏzG2*O6.$_vyޝEh/z翽?X|_kEV(ݟz.55w?9uIR U sRcp7[g9ŚsJ6XM@F<5a*W2;./f+p:EI= ˋ^^4l `EcbH鮼h{o\^tYlr 4FyxWp0uE:{GSN,/:8塚h  4byѶH<l0)*qhID u C40:!]QdKͨ-+4&KaJ#cF+W}+- &ôoR(@rZ4)^kIPWwckCw %A I_fѡBc%,A{w{l_5loo Ȕ>9=y&2[wTޕ6r$Bed)K$& Jt[%vsu5N,gꩳuZ+l7Z"(%}>?L*E RK,0E<񎻢)y.$6AVY+s;訣#V#֠޽ж=ݬ I? (;3Hd\w‹˘?O~ f״s?JZZF39GWWd2joqs (L &htА"d't*ԁIq P MӐvCV ڄ`.^$(, 28O?,Y%|J@ZNS+XZYcz bu%GTQ4R-Dc@XV;,( ut EJSu9$"G& R+UmA:Zvʶ 3ֶiОmA]?j m 3mAZSTZ y6|BֵRXc$fG %6=ڂʇ"'|?Lɔ:7gm?_u?_GhgC*, uJ=$=P]Wcड़I5 6؎rHe~ImdT%΃%"Oy}?pJAgBjƳQ= z8IVR+ZT&; TP{VՖX D\NZtTe)p9DEdQ"TBU%SֲWiTk0^MqZ 9OF)󃇱<}=J@7{'P'TK\5Jd7L/|K>G2 7JeeBEbK;>yD9̗ 3Cwbη]I3#i1cwu:=f]ޛsr}{AM&qJ&ޤH4hBhI^~mE#0߸h:TQ c} c>Qg,JbX$eoH'f@՗_sDt=2}]g++6 yO\,>[n#T^?6bB)_!@s:jQ$P\F&EYMN3&?*e+kAX oϟll6DO!"EW ]p C6j{#D!zH)exΖe02Fx~..>~>{- 3`</"%˄"9E"ž bPY(.(}@FU\"#y@Dg}*U) RzEA y26sz+x$y~2NG]1hkǧϬzA2 ptsAv0ZmƓ;ƭڈe(E{AOk׸];[Sqɕ> d՗VrF$HI 0FA:/X X6Z%y~r=LHiٻV4~,Xi'F4K`mJfEU~aǴDY/P!w>#>'z` -vVUp= Jq{ש`'Y52HhK}q 6&NL >2IUtrp.l ᶑn?G#RicA&v m%( -]{a g[MVk#Yɪ,ّm-)+PZ0Ռ`/ũ S{LMp};h[3zt4#TsX C.YҤ8H4|S4x>MGn29;@TNL^ڕBo|Sy1-lruKK{СeVMG lOW /bH*-C>%Tז N߼)_"Xpͥyԟ.?MC'rI,B2qVZtfZP=8:TٙVHmA_iFAGH2'tn UFR*5%z8)p\71IL/9mSF39GW1*M7}_~q5.w9B9"Y˘IgM =ƆA">{m 45gPFBEX!|> h苎G񿝌ۛ2,oC0eھs4:/mC8K`fRPj8EfL*):(^)G(]%Yj"{ό2.bBpc <=E`^sF4{8q$*(j'![.kց$(zob2kϛc"{N}"JV)z>]WXؾȭ'Ӳ TRtyhZ g4_^.ӳˇ﬐W͞gFKN-N|4Kb]Gߤ4~(+&ǤcNl椡;JVdζ}<}6{oNdpU> PTXC?m* Veͷ*VJ ztus{x֩FїV]fB̜؈c?zW;GxsP:;GyFK}E΄ޚȻฺ^N_wJMW`0Bh#zuv7?rZҨAܶF'չ⫯4<{alVZ/WbTÅFAOJ0|> 1|`>@ fM fa jHPχdJI~Tk:w3'KT\z ޭ_jos+y}S,߶eT1Mn㬺ܪz0?_/u>~0ua*MڏL;cN;'kƸU?hkPmgVNq0 }[306 6 Oבd+q zXpʄ c+h_|2B . vmkǹ}ɥ*kxi3>3|esx#] P+~hZ|Yhs3)7A!|y'{Skм>+uh5{VdLjM̽H* 3)!-ά1S 9er&Mጸy#s'ڡevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004031146315156372424017710 0ustar rootrootMar 18 00:08:00 crc systemd[1]: Starting Kubernetes Kubelet... Mar 18 00:08:00 crc restorecon[4683]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:00 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 00:08:01 crc restorecon[4683]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 00:08:01 crc restorecon[4683]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 18 00:08:02 crc kubenswrapper[4983]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 18 00:08:02 crc kubenswrapper[4983]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 18 00:08:02 crc kubenswrapper[4983]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 18 00:08:02 crc kubenswrapper[4983]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 18 00:08:02 crc kubenswrapper[4983]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 18 00:08:02 crc kubenswrapper[4983]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.543188 4983 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551041 4983 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551077 4983 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551092 4983 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551103 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551113 4983 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551124 4983 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551133 4983 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551142 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551150 4983 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551159 4983 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551166 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551174 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551182 4983 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551190 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551198 4983 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551206 4983 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551214 4983 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551222 4983 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551231 4983 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551239 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551248 4983 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551255 4983 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551264 4983 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551272 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551280 4983 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551289 4983 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551300 4983 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551312 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551323 4983 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551333 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551342 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551376 4983 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551386 4983 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551396 4983 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551404 4983 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551412 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551420 4983 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551429 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551437 4983 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551445 4983 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551452 4983 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551461 4983 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551470 4983 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551478 4983 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551490 4983 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551499 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551509 4983 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551517 4983 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551525 4983 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551533 4983 feature_gate.go:330] unrecognized feature gate: Example Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551541 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551550 4983 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551558 4983 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551566 4983 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551574 4983 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551581 4983 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551589 4983 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551597 4983 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551606 4983 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551613 4983 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551623 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551630 4983 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551638 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551650 4983 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551659 4983 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551668 4983 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551676 4983 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551684 4983 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551691 4983 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551699 4983 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.551707 4983 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.551909 4983 flags.go:64] FLAG: --address="0.0.0.0" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.551951 4983 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.551969 4983 flags.go:64] FLAG: --anonymous-auth="true" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.551980 4983 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.551992 4983 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552002 4983 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552014 4983 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552026 4983 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552035 4983 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552045 4983 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552055 4983 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552064 4983 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552073 4983 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552082 4983 flags.go:64] FLAG: --cgroup-root="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552091 4983 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552100 4983 flags.go:64] FLAG: --client-ca-file="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552108 4983 flags.go:64] FLAG: --cloud-config="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552117 4983 flags.go:64] FLAG: --cloud-provider="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552128 4983 flags.go:64] FLAG: --cluster-dns="[]" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552139 4983 flags.go:64] FLAG: --cluster-domain="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552148 4983 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552157 4983 flags.go:64] FLAG: --config-dir="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552166 4983 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552176 4983 flags.go:64] FLAG: --container-log-max-files="5" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552187 4983 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552196 4983 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552206 4983 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552216 4983 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552225 4983 flags.go:64] FLAG: --contention-profiling="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552249 4983 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552259 4983 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552269 4983 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552277 4983 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552289 4983 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552298 4983 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552306 4983 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552315 4983 flags.go:64] FLAG: --enable-load-reader="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552324 4983 flags.go:64] FLAG: --enable-server="true" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552333 4983 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552345 4983 flags.go:64] FLAG: --event-burst="100" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552354 4983 flags.go:64] FLAG: --event-qps="50" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552363 4983 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552373 4983 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552383 4983 flags.go:64] FLAG: --eviction-hard="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552394 4983 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552403 4983 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552413 4983 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552423 4983 flags.go:64] FLAG: --eviction-soft="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552434 4983 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552444 4983 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552457 4983 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552468 4983 flags.go:64] FLAG: --experimental-mounter-path="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552478 4983 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552488 4983 flags.go:64] FLAG: --fail-swap-on="true" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552499 4983 flags.go:64] FLAG: --feature-gates="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552511 4983 flags.go:64] FLAG: --file-check-frequency="20s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552523 4983 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552533 4983 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552544 4983 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552554 4983 flags.go:64] FLAG: --healthz-port="10248" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552565 4983 flags.go:64] FLAG: --help="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552575 4983 flags.go:64] FLAG: --hostname-override="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552586 4983 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552596 4983 flags.go:64] FLAG: --http-check-frequency="20s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552607 4983 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552617 4983 flags.go:64] FLAG: --image-credential-provider-config="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552627 4983 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552639 4983 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552650 4983 flags.go:64] FLAG: --image-service-endpoint="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552660 4983 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552671 4983 flags.go:64] FLAG: --kube-api-burst="100" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552682 4983 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552694 4983 flags.go:64] FLAG: --kube-api-qps="50" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552703 4983 flags.go:64] FLAG: --kube-reserved="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552713 4983 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552722 4983 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552732 4983 flags.go:64] FLAG: --kubelet-cgroups="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552741 4983 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552750 4983 flags.go:64] FLAG: --lock-file="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552760 4983 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552770 4983 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552780 4983 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552803 4983 flags.go:64] FLAG: --log-json-split-stream="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552814 4983 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552823 4983 flags.go:64] FLAG: --log-text-split-stream="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552833 4983 flags.go:64] FLAG: --logging-format="text" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552842 4983 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552852 4983 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552862 4983 flags.go:64] FLAG: --manifest-url="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552871 4983 flags.go:64] FLAG: --manifest-url-header="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552883 4983 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552893 4983 flags.go:64] FLAG: --max-open-files="1000000" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552904 4983 flags.go:64] FLAG: --max-pods="110" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552914 4983 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552923 4983 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552955 4983 flags.go:64] FLAG: --memory-manager-policy="None" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552965 4983 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552974 4983 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552985 4983 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.552994 4983 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553014 4983 flags.go:64] FLAG: --node-status-max-images="50" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553024 4983 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553033 4983 flags.go:64] FLAG: --oom-score-adj="-999" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553043 4983 flags.go:64] FLAG: --pod-cidr="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553053 4983 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553066 4983 flags.go:64] FLAG: --pod-manifest-path="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553075 4983 flags.go:64] FLAG: --pod-max-pids="-1" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553085 4983 flags.go:64] FLAG: --pods-per-core="0" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553094 4983 flags.go:64] FLAG: --port="10250" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553103 4983 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553112 4983 flags.go:64] FLAG: --provider-id="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553121 4983 flags.go:64] FLAG: --qos-reserved="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553130 4983 flags.go:64] FLAG: --read-only-port="10255" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553139 4983 flags.go:64] FLAG: --register-node="true" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553149 4983 flags.go:64] FLAG: --register-schedulable="true" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553161 4983 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553176 4983 flags.go:64] FLAG: --registry-burst="10" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553185 4983 flags.go:64] FLAG: --registry-qps="5" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553194 4983 flags.go:64] FLAG: --reserved-cpus="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553203 4983 flags.go:64] FLAG: --reserved-memory="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553213 4983 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553223 4983 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553232 4983 flags.go:64] FLAG: --rotate-certificates="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553241 4983 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553250 4983 flags.go:64] FLAG: --runonce="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553259 4983 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553268 4983 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553278 4983 flags.go:64] FLAG: --seccomp-default="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553287 4983 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553297 4983 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553306 4983 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553315 4983 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553324 4983 flags.go:64] FLAG: --storage-driver-password="root" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553333 4983 flags.go:64] FLAG: --storage-driver-secure="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553342 4983 flags.go:64] FLAG: --storage-driver-table="stats" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553352 4983 flags.go:64] FLAG: --storage-driver-user="root" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553360 4983 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553370 4983 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553380 4983 flags.go:64] FLAG: --system-cgroups="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553389 4983 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553403 4983 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553413 4983 flags.go:64] FLAG: --tls-cert-file="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553424 4983 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553436 4983 flags.go:64] FLAG: --tls-min-version="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553446 4983 flags.go:64] FLAG: --tls-private-key-file="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553457 4983 flags.go:64] FLAG: --topology-manager-policy="none" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553467 4983 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553478 4983 flags.go:64] FLAG: --topology-manager-scope="container" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553489 4983 flags.go:64] FLAG: --v="2" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553502 4983 flags.go:64] FLAG: --version="false" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553515 4983 flags.go:64] FLAG: --vmodule="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553528 4983 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.553539 4983 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553779 4983 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553791 4983 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553801 4983 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553813 4983 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553827 4983 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553839 4983 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553850 4983 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553860 4983 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553904 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553917 4983 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553959 4983 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553969 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553979 4983 feature_gate.go:330] unrecognized feature gate: Example Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.553989 4983 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554000 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554009 4983 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554019 4983 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554027 4983 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554035 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554043 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554051 4983 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554058 4983 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554066 4983 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554074 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554082 4983 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554096 4983 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554104 4983 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554112 4983 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554119 4983 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554127 4983 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554137 4983 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554146 4983 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554155 4983 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554164 4983 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554174 4983 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554182 4983 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554190 4983 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554198 4983 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554206 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554214 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554222 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554230 4983 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554237 4983 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554245 4983 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554256 4983 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554264 4983 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554273 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554281 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554290 4983 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554298 4983 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554305 4983 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554313 4983 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554321 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554329 4983 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554337 4983 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554345 4983 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554353 4983 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554363 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554371 4983 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554379 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554388 4983 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554395 4983 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554403 4983 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554411 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554418 4983 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554426 4983 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554434 4983 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554441 4983 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554449 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554457 4983 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.554466 4983 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.554488 4983 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.566329 4983 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.566398 4983 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566552 4983 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566573 4983 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566583 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566593 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566603 4983 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566613 4983 feature_gate.go:330] unrecognized feature gate: Example Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566621 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566632 4983 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566643 4983 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566652 4983 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566661 4983 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566669 4983 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566678 4983 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566687 4983 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566695 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566703 4983 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566711 4983 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566718 4983 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566726 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566734 4983 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566741 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566749 4983 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566757 4983 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566799 4983 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566809 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566820 4983 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566829 4983 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566837 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566846 4983 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566854 4983 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566863 4983 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566871 4983 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566879 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566887 4983 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566896 4983 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566903 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566911 4983 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566919 4983 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566950 4983 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566958 4983 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566966 4983 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566974 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566982 4983 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566991 4983 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.566998 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567006 4983 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567013 4983 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567021 4983 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567029 4983 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567037 4983 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567045 4983 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567053 4983 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567060 4983 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567068 4983 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567076 4983 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567084 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567091 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567100 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567107 4983 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567116 4983 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567123 4983 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567132 4983 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567140 4983 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567148 4983 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567157 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567165 4983 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567175 4983 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567186 4983 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567197 4983 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567207 4983 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567215 4983 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.567230 4983 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567523 4983 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567536 4983 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567545 4983 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567553 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567562 4983 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567571 4983 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567579 4983 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567590 4983 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567599 4983 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567607 4983 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567615 4983 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567624 4983 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567631 4983 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567639 4983 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567647 4983 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567655 4983 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567663 4983 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567674 4983 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567685 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567694 4983 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567703 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567712 4983 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567721 4983 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567732 4983 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567742 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567752 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567761 4983 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567770 4983 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567779 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567788 4983 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567798 4983 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567809 4983 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567817 4983 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567826 4983 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567835 4983 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567844 4983 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567851 4983 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567859 4983 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567866 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567874 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567882 4983 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567890 4983 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567898 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567906 4983 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567913 4983 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567921 4983 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567969 4983 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.567987 4983 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568000 4983 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568011 4983 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568024 4983 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568035 4983 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568045 4983 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568055 4983 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568064 4983 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568072 4983 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568080 4983 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568088 4983 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568096 4983 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568104 4983 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568111 4983 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568124 4983 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568131 4983 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568139 4983 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568147 4983 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568155 4983 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568163 4983 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568171 4983 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568179 4983 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568186 4983 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.568194 4983 feature_gate.go:330] unrecognized feature gate: Example Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.568206 4983 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.569968 4983 server.go:940] "Client rotation is on, will bootstrap in background" Mar 18 00:08:02 crc kubenswrapper[4983]: E0318 00:08:02.575510 4983 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.580345 4983 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.580523 4983 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.582833 4983 server.go:997] "Starting client certificate rotation" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.582882 4983 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.583997 4983 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.609786 4983 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.612572 4983 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 18 00:08:02 crc kubenswrapper[4983]: E0318 00:08:02.613054 4983 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.632455 4983 log.go:25] "Validated CRI v1 runtime API" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.673916 4983 log.go:25] "Validated CRI v1 image API" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.683210 4983 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.690559 4983 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-18-00-03-18-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.690606 4983 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.720871 4983 manager.go:217] Machine: {Timestamp:2026-03-18 00:08:02.717903925 +0000 UTC m=+0.615278605 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:3cf496ed-375e-4961-a03c-4bc8923d9044 BootID:20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:c0:21:ed Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:c0:21:ed Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:db:db:8d Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:85:ec:63 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:92:72:6b Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:68:29:30 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:1a:9f:3d:54:d2:eb Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:a2:28:f3:4a:28:fa Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.721406 4983 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.721717 4983 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.724358 4983 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.724754 4983 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.724816 4983 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.725182 4983 topology_manager.go:138] "Creating topology manager with none policy" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.725200 4983 container_manager_linux.go:303] "Creating device plugin manager" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.725839 4983 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.725890 4983 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.727017 4983 state_mem.go:36] "Initialized new in-memory state store" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.727149 4983 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.731841 4983 kubelet.go:418] "Attempting to sync node with API server" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.731875 4983 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.731902 4983 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.731945 4983 kubelet.go:324] "Adding apiserver pod source" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.731965 4983 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.739041 4983 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.741562 4983 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.741817 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.741818 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Mar 18 00:08:02 crc kubenswrapper[4983]: E0318 00:08:02.741959 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Mar 18 00:08:02 crc kubenswrapper[4983]: E0318 00:08:02.741980 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.744439 4983 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.746736 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.746775 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.746789 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.746804 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.746825 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.746839 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.746856 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.746877 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.746893 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.746907 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.746957 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.746971 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.748667 4983 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.749501 4983 server.go:1280] "Started kubelet" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.749705 4983 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.749819 4983 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.750673 4983 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 18 00:08:02 crc systemd[1]: Started Kubernetes Kubelet. Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.752839 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.754038 4983 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.761428 4983 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.763002 4983 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.763036 4983 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 18 00:08:02 crc kubenswrapper[4983]: E0318 00:08:02.763252 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.763604 4983 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.765195 4983 factory.go:55] Registering systemd factory Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.765244 4983 factory.go:221] Registration of the systemd container factory successfully Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.765459 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Mar 18 00:08:02 crc kubenswrapper[4983]: E0318 00:08:02.765555 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.766774 4983 server.go:460] "Adding debug handlers to kubelet server" Mar 18 00:08:02 crc kubenswrapper[4983]: E0318 00:08:02.765846 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="200ms" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.768167 4983 factory.go:153] Registering CRI-O factory Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.768205 4983 factory.go:221] Registration of the crio container factory successfully Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.768317 4983 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.768364 4983 factory.go:103] Registering Raw factory Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.768403 4983 manager.go:1196] Started watching for new ooms in manager Mar 18 00:08:02 crc kubenswrapper[4983]: E0318 00:08:02.767000 4983 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189dc6dd778a15d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.749453777 +0000 UTC m=+0.646828417,LastTimestamp:2026-03-18 00:08:02.749453777 +0000 UTC m=+0.646828417,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.779656 4983 manager.go:319] Starting recovery of all containers Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.783825 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.783897 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.783921 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.783967 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.783987 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784008 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784028 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784049 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784071 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784091 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784130 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784153 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784172 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784196 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784230 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784252 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784272 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784290 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784309 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784359 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784379 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784419 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784440 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784460 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784481 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784501 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784526 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784546 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784565 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784586 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784605 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784628 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784647 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784665 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784682 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784706 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784724 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784748 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784767 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784786 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784804 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784823 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784844 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784862 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784884 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784904 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784922 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784976 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.784998 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785018 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785036 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785054 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785078 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785101 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785121 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785140 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785161 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785180 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785199 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785221 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785239 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785257 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785276 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785295 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785313 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785333 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785352 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785370 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785389 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785408 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785427 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785456 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785476 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785494 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785513 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785532 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785551 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785569 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785589 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785610 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785628 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785648 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785665 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785684 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785703 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785721 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785739 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785758 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785777 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785796 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785815 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.785834 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789187 4983 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789232 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789255 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789277 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789296 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789316 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789337 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789375 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789394 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789415 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789436 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789455 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789473 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789498 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789519 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789539 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789560 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789583 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789604 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789624 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789644 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789664 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789683 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789704 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789723 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789741 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789761 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789782 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789799 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789818 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789835 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789852 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789870 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789887 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789904 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789923 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789967 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.789985 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790003 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790021 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790039 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790058 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790074 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790092 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790109 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790129 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790146 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790201 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790218 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790238 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790256 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790275 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790293 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790311 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790356 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790376 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790392 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790410 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790431 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790457 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790545 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790564 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790582 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790601 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790619 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790636 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790655 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790676 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790694 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790733 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790754 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790773 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790795 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790815 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790834 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790853 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790874 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790895 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790916 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790959 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790978 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.790996 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791016 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791036 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791055 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791073 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791091 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791110 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791127 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791146 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791164 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791182 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791200 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791220 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791239 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791260 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791278 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791295 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791313 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791332 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791351 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791370 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791387 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791405 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791425 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791443 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791462 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791480 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791498 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791516 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791534 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791553 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791574 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791594 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791613 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791633 4983 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791709 4983 reconstruct.go:97] "Volume reconstruction finished" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.791723 4983 reconciler.go:26] "Reconciler: start to sync state" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.810198 4983 manager.go:324] Recovery completed Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.830023 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.832239 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.832302 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.832319 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.833522 4983 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.833559 4983 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.833604 4983 state_mem.go:36] "Initialized new in-memory state store" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.839788 4983 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.843114 4983 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.843181 4983 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.843223 4983 kubelet.go:2335] "Starting kubelet main sync loop" Mar 18 00:08:02 crc kubenswrapper[4983]: E0318 00:08:02.843402 4983 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 18 00:08:02 crc kubenswrapper[4983]: W0318 00:08:02.844025 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Mar 18 00:08:02 crc kubenswrapper[4983]: E0318 00:08:02.844088 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.848400 4983 policy_none.go:49] "None policy: Start" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.851157 4983 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.851206 4983 state_mem.go:35] "Initializing new in-memory state store" Mar 18 00:08:02 crc kubenswrapper[4983]: E0318 00:08:02.863724 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.912383 4983 manager.go:334] "Starting Device Plugin manager" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.912532 4983 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.912560 4983 server.go:79] "Starting device plugin registration server" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.913431 4983 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.913468 4983 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.913891 4983 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.914059 4983 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.914074 4983 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 18 00:08:02 crc kubenswrapper[4983]: E0318 00:08:02.926814 4983 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.943738 4983 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.943971 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.945807 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.945866 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.945889 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.946233 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.946776 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.947008 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.947484 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.947523 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.947538 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.947854 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.948894 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.949172 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.949023 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.949430 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.949470 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.950162 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.950201 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.950216 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.950358 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.950568 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.950668 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.951413 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.951478 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.951498 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.951579 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.951601 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.951613 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.951741 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.951956 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.952024 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.952236 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.952299 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.952325 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.953306 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.953351 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.953369 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.953543 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.953587 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.953601 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.953625 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.953631 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.954832 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.954880 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.954897 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:02 crc kubenswrapper[4983]: E0318 00:08:02.968478 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="400ms" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.994618 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.994736 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.994819 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.994900 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.995067 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.995163 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.995242 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.995268 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.995392 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.995435 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.995496 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.995557 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.995607 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.995676 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 00:08:02 crc kubenswrapper[4983]: I0318 00:08:02.995746 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.015986 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.018974 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.019034 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.019049 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.019092 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:08:03 crc kubenswrapper[4983]: E0318 00:08:03.019813 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097050 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097380 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097506 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097572 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097639 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097651 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097702 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097704 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097737 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097831 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097780 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097792 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097889 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097954 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097990 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098028 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098042 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.097968 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098097 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098109 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098065 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098064 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098207 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098257 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098289 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098306 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098321 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098349 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098452 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.098481 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.220665 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.222548 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.222623 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.222639 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.222679 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:08:03 crc kubenswrapper[4983]: E0318 00:08:03.223233 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.294293 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.307462 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.325200 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: W0318 00:08:03.349363 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-f97e3eb252469cf58faaacd2b8675b51ac4b5e802760b6f10138986d854084e4 WatchSource:0}: Error finding container f97e3eb252469cf58faaacd2b8675b51ac4b5e802760b6f10138986d854084e4: Status 404 returned error can't find the container with id f97e3eb252469cf58faaacd2b8675b51ac4b5e802760b6f10138986d854084e4 Mar 18 00:08:03 crc kubenswrapper[4983]: W0318 00:08:03.352514 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-26be0d574ef2a74f0fd6cb35be8a2d2625fff4dd898fa5a06c0860849cd753c4 WatchSource:0}: Error finding container 26be0d574ef2a74f0fd6cb35be8a2d2625fff4dd898fa5a06c0860849cd753c4: Status 404 returned error can't find the container with id 26be0d574ef2a74f0fd6cb35be8a2d2625fff4dd898fa5a06c0860849cd753c4 Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.355757 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: W0318 00:08:03.356577 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-bce2eff2b6b237a26ff261a19f92c044131ae4d689738cb14c6acd75ba16272e WatchSource:0}: Error finding container bce2eff2b6b237a26ff261a19f92c044131ae4d689738cb14c6acd75ba16272e: Status 404 returned error can't find the container with id bce2eff2b6b237a26ff261a19f92c044131ae4d689738cb14c6acd75ba16272e Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.366131 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 00:08:03 crc kubenswrapper[4983]: E0318 00:08:03.370124 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="800ms" Mar 18 00:08:03 crc kubenswrapper[4983]: W0318 00:08:03.371154 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b0c3c60e5b6eb46dbdcab5dca83acdf39cdc826b438625614577a66a3a8a25c4 WatchSource:0}: Error finding container b0c3c60e5b6eb46dbdcab5dca83acdf39cdc826b438625614577a66a3a8a25c4: Status 404 returned error can't find the container with id b0c3c60e5b6eb46dbdcab5dca83acdf39cdc826b438625614577a66a3a8a25c4 Mar 18 00:08:03 crc kubenswrapper[4983]: W0318 00:08:03.406170 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-1a64577e2a98d1c3ff81372dc30cf8673bf79212e89234e4ada1f7abf47d600c WatchSource:0}: Error finding container 1a64577e2a98d1c3ff81372dc30cf8673bf79212e89234e4ada1f7abf47d600c: Status 404 returned error can't find the container with id 1a64577e2a98d1c3ff81372dc30cf8673bf79212e89234e4ada1f7abf47d600c Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.624297 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.625734 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.625798 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.625817 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.625861 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:08:03 crc kubenswrapper[4983]: E0318 00:08:03.626481 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.754756 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.851507 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bce2eff2b6b237a26ff261a19f92c044131ae4d689738cb14c6acd75ba16272e"} Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.852574 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"26be0d574ef2a74f0fd6cb35be8a2d2625fff4dd898fa5a06c0860849cd753c4"} Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.853388 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f97e3eb252469cf58faaacd2b8675b51ac4b5e802760b6f10138986d854084e4"} Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.854165 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1a64577e2a98d1c3ff81372dc30cf8673bf79212e89234e4ada1f7abf47d600c"} Mar 18 00:08:03 crc kubenswrapper[4983]: I0318 00:08:03.855439 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b0c3c60e5b6eb46dbdcab5dca83acdf39cdc826b438625614577a66a3a8a25c4"} Mar 18 00:08:04 crc kubenswrapper[4983]: W0318 00:08:04.007005 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Mar 18 00:08:04 crc kubenswrapper[4983]: E0318 00:08:04.007618 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Mar 18 00:08:04 crc kubenswrapper[4983]: W0318 00:08:04.164971 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Mar 18 00:08:04 crc kubenswrapper[4983]: E0318 00:08:04.165115 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Mar 18 00:08:04 crc kubenswrapper[4983]: E0318 00:08:04.170796 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="1.6s" Mar 18 00:08:04 crc kubenswrapper[4983]: W0318 00:08:04.333646 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Mar 18 00:08:04 crc kubenswrapper[4983]: E0318 00:08:04.333787 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Mar 18 00:08:04 crc kubenswrapper[4983]: W0318 00:08:04.409918 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Mar 18 00:08:04 crc kubenswrapper[4983]: E0318 00:08:04.410086 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.427218 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.429236 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.429308 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.429329 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.429370 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:08:04 crc kubenswrapper[4983]: E0318 00:08:04.430116 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.704515 4983 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 00:08:04 crc kubenswrapper[4983]: E0318 00:08:04.705492 4983 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.753630 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.860256 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26"} Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.860298 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d"} Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.860312 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2"} Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.862003 4983 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716" exitCode=0 Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.862057 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716"} Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.862215 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.863712 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.863756 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.863765 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.865016 4983 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9" exitCode=0 Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.865067 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9"} Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.865166 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.865371 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.866465 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.866486 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.866495 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.866840 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.866856 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.866863 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.869361 4983 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="655e9f142d07c3af06937170bb4eaa3ec5fbbb608685854cc1545665aec22b24" exitCode=0 Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.869407 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"655e9f142d07c3af06937170bb4eaa3ec5fbbb608685854cc1545665aec22b24"} Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.869465 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.870310 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.870325 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.870332 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.872750 4983 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88" exitCode=0 Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.872783 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88"} Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.872854 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.873467 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.873484 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:04 crc kubenswrapper[4983]: I0318 00:08:04.873492 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.754034 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Mar 18 00:08:05 crc kubenswrapper[4983]: E0318 00:08:05.772012 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="3.2s" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.885018 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0befc5603e720538a3a0e9c40a50792f3e8d07c61929b87bd424427be0e54ddb"} Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.885137 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.888699 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.888785 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.888811 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.892535 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d3939424d6655bea60e9bad967f755a313411e8a699563cd3a9ef2b2e2354362"} Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.892570 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d84f4ccd83826ef23c1583cae04ee8092f94bb5a2f9c948f97e503a9b1858201"} Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.892579 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"81167a1f7781d69d700f7647928cbb027a40efec3129a860f62f82b12e2d9656"} Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.892696 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.893746 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.893774 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.893783 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.900827 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b"} Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.901093 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.903043 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.903102 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.903126 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.908914 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256"} Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.908990 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38"} Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.909013 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349"} Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.909031 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816"} Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.911775 4983 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163" exitCode=0 Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.911815 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163"} Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.911987 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.913681 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.913711 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:05 crc kubenswrapper[4983]: I0318 00:08:05.913723 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:05 crc kubenswrapper[4983]: W0318 00:08:05.941354 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Mar 18 00:08:05 crc kubenswrapper[4983]: E0318 00:08:05.941476 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.031244 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.032923 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.033051 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.033071 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.033108 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:08:06 crc kubenswrapper[4983]: E0318 00:08:06.033828 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.078941 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.921564 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"350b6156522c018873b37e48f03b0b90eae9d690480d8f4bcf44e1c751443bae"} Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.921660 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.923026 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.923065 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.923084 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.925494 4983 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb" exitCode=0 Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.925628 4983 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.925648 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.925678 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.925708 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.925770 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb"} Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.925818 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.930682 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.930728 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.930743 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.930792 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.930799 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.931124 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.931147 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.930845 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.931222 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.931240 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.931295 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:06 crc kubenswrapper[4983]: I0318 00:08:06.932349 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:07 crc kubenswrapper[4983]: I0318 00:08:07.934378 4983 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 18 00:08:07 crc kubenswrapper[4983]: I0318 00:08:07.934444 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:07 crc kubenswrapper[4983]: I0318 00:08:07.935210 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6"} Mar 18 00:08:07 crc kubenswrapper[4983]: I0318 00:08:07.935292 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13"} Mar 18 00:08:07 crc kubenswrapper[4983]: I0318 00:08:07.935321 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810"} Mar 18 00:08:07 crc kubenswrapper[4983]: I0318 00:08:07.935391 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:07 crc kubenswrapper[4983]: I0318 00:08:07.935686 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:07 crc kubenswrapper[4983]: I0318 00:08:07.935741 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:07 crc kubenswrapper[4983]: I0318 00:08:07.935767 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:07 crc kubenswrapper[4983]: I0318 00:08:07.936711 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:07 crc kubenswrapper[4983]: I0318 00:08:07.936767 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:07 crc kubenswrapper[4983]: I0318 00:08:07.936789 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:08 crc kubenswrapper[4983]: I0318 00:08:08.116955 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 00:08:08 crc kubenswrapper[4983]: I0318 00:08:08.117216 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:08 crc kubenswrapper[4983]: I0318 00:08:08.118798 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:08 crc kubenswrapper[4983]: I0318 00:08:08.118888 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:08 crc kubenswrapper[4983]: I0318 00:08:08.118922 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:08 crc kubenswrapper[4983]: I0318 00:08:08.945004 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2"} Mar 18 00:08:08 crc kubenswrapper[4983]: I0318 00:08:08.945068 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0"} Mar 18 00:08:08 crc kubenswrapper[4983]: I0318 00:08:08.945267 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:08 crc kubenswrapper[4983]: I0318 00:08:08.946610 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:08 crc kubenswrapper[4983]: I0318 00:08:08.946654 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:08 crc kubenswrapper[4983]: I0318 00:08:08.946671 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:08 crc kubenswrapper[4983]: I0318 00:08:08.987159 4983 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.234596 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.236480 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.236728 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.236877 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.237136 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.373756 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.373984 4983 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.374054 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.375555 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.375637 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.375664 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.947740 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.948966 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.949005 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:09 crc kubenswrapper[4983]: I0318 00:08:09.949019 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:10 crc kubenswrapper[4983]: I0318 00:08:10.820295 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:10 crc kubenswrapper[4983]: I0318 00:08:10.820575 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:10 crc kubenswrapper[4983]: I0318 00:08:10.822141 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:10 crc kubenswrapper[4983]: I0318 00:08:10.822217 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:10 crc kubenswrapper[4983]: I0318 00:08:10.822241 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:11 crc kubenswrapper[4983]: I0318 00:08:11.082954 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:11 crc kubenswrapper[4983]: I0318 00:08:11.083262 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:11 crc kubenswrapper[4983]: I0318 00:08:11.084879 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:11 crc kubenswrapper[4983]: I0318 00:08:11.084922 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:11 crc kubenswrapper[4983]: I0318 00:08:11.085007 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:11 crc kubenswrapper[4983]: I0318 00:08:11.633034 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 18 00:08:11 crc kubenswrapper[4983]: I0318 00:08:11.633273 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:11 crc kubenswrapper[4983]: I0318 00:08:11.634821 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:11 crc kubenswrapper[4983]: I0318 00:08:11.634896 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:11 crc kubenswrapper[4983]: I0318 00:08:11.634984 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.162048 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.162335 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.163824 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.163886 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.163903 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.821635 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.822421 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.824071 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.824134 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.824159 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.830134 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:12 crc kubenswrapper[4983]: E0318 00:08:12.927190 4983 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.955706 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.955778 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.957206 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.957260 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:12 crc kubenswrapper[4983]: I0318 00:08:12.957280 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:13 crc kubenswrapper[4983]: I0318 00:08:13.959218 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:13 crc kubenswrapper[4983]: I0318 00:08:13.960621 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:13 crc kubenswrapper[4983]: I0318 00:08:13.960868 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:13 crc kubenswrapper[4983]: I0318 00:08:13.961068 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:13 crc kubenswrapper[4983]: I0318 00:08:13.967469 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:14 crc kubenswrapper[4983]: I0318 00:08:14.963525 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:14 crc kubenswrapper[4983]: I0318 00:08:14.966145 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:14 crc kubenswrapper[4983]: I0318 00:08:14.966338 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:14 crc kubenswrapper[4983]: I0318 00:08:14.966422 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:15 crc kubenswrapper[4983]: I0318 00:08:15.350147 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:15 crc kubenswrapper[4983]: I0318 00:08:15.965278 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:15 crc kubenswrapper[4983]: I0318 00:08:15.966765 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:15 crc kubenswrapper[4983]: I0318 00:08:15.966810 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:15 crc kubenswrapper[4983]: I0318 00:08:15.966831 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:16 crc kubenswrapper[4983]: W0318 00:08:16.645723 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 18 00:08:16 crc kubenswrapper[4983]: I0318 00:08:16.645880 4983 trace.go:236] Trace[80312067]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Mar-2026 00:08:06.643) (total time: 10002ms): Mar 18 00:08:16 crc kubenswrapper[4983]: Trace[80312067]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (00:08:16.645) Mar 18 00:08:16 crc kubenswrapper[4983]: Trace[80312067]: [10.002054157s] [10.002054157s] END Mar 18 00:08:16 crc kubenswrapper[4983]: E0318 00:08:16.645922 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 18 00:08:16 crc kubenswrapper[4983]: I0318 00:08:16.755272 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 18 00:08:17 crc kubenswrapper[4983]: W0318 00:08:17.050131 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.050975 4983 trace.go:236] Trace[1691539739]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Mar-2026 00:08:07.044) (total time: 10006ms): Mar 18 00:08:17 crc kubenswrapper[4983]: Trace[1691539739]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10005ms (00:08:17.050) Mar 18 00:08:17 crc kubenswrapper[4983]: Trace[1691539739]: [10.006402026s] [10.006402026s] END Mar 18 00:08:17 crc kubenswrapper[4983]: E0318 00:08:17.051195 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 18 00:08:17 crc kubenswrapper[4983]: W0318 00:08:17.071064 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.071233 4983 trace.go:236] Trace[1366930760]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Mar-2026 00:08:07.069) (total time: 10001ms): Mar 18 00:08:17 crc kubenswrapper[4983]: Trace[1366930760]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (00:08:17.071) Mar 18 00:08:17 crc kubenswrapper[4983]: Trace[1366930760]: [10.001703584s] [10.001703584s] END Mar 18 00:08:17 crc kubenswrapper[4983]: E0318 00:08:17.071393 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 18 00:08:17 crc kubenswrapper[4983]: E0318 00:08:17.267464 4983 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:17Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189dc6dd778a15d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.749453777 +0000 UTC m=+0.646828417,LastTimestamp:2026-03-18 00:08:02.749453777 +0000 UTC m=+0.646828417,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:17 crc kubenswrapper[4983]: W0318 00:08:17.269421 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:17Z is after 2026-02-23T05:33:13Z Mar 18 00:08:17 crc kubenswrapper[4983]: E0318 00:08:17.269846 4983 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:17Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:17 crc kubenswrapper[4983]: E0318 00:08:17.269840 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:17Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:17 crc kubenswrapper[4983]: E0318 00:08:17.270835 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:17Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 18 00:08:17 crc kubenswrapper[4983]: E0318 00:08:17.272323 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:17Z is after 2026-02-23T05:33:13Z" node="crc" Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.274776 4983 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.274865 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.280408 4983 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.280779 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.364367 4983 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.364450 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.758820 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:17Z is after 2026-02-23T05:33:13Z Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.972326 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.974654 4983 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="350b6156522c018873b37e48f03b0b90eae9d690480d8f4bcf44e1c751443bae" exitCode=255 Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.974718 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"350b6156522c018873b37e48f03b0b90eae9d690480d8f4bcf44e1c751443bae"} Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.974973 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.975856 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.975910 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.975969 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:17 crc kubenswrapper[4983]: I0318 00:08:17.976713 4983 scope.go:117] "RemoveContainer" containerID="350b6156522c018873b37e48f03b0b90eae9d690480d8f4bcf44e1c751443bae" Mar 18 00:08:18 crc kubenswrapper[4983]: I0318 00:08:18.350727 4983 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 00:08:18 crc kubenswrapper[4983]: I0318 00:08:18.350799 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 00:08:18 crc kubenswrapper[4983]: I0318 00:08:18.756949 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:18Z is after 2026-02-23T05:33:13Z Mar 18 00:08:18 crc kubenswrapper[4983]: I0318 00:08:18.980623 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 18 00:08:18 crc kubenswrapper[4983]: I0318 00:08:18.984998 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"772ad4d5c7862487c75cca409c63a8bf15f47d3b12a2cc71e1798e32691d7424"} Mar 18 00:08:18 crc kubenswrapper[4983]: I0318 00:08:18.985271 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:18 crc kubenswrapper[4983]: I0318 00:08:18.986614 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:18 crc kubenswrapper[4983]: I0318 00:08:18.986694 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:18 crc kubenswrapper[4983]: I0318 00:08:18.986719 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:19 crc kubenswrapper[4983]: I0318 00:08:19.760607 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:19Z is after 2026-02-23T05:33:13Z Mar 18 00:08:19 crc kubenswrapper[4983]: I0318 00:08:19.990823 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 18 00:08:19 crc kubenswrapper[4983]: I0318 00:08:19.991567 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 18 00:08:19 crc kubenswrapper[4983]: I0318 00:08:19.993694 4983 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="772ad4d5c7862487c75cca409c63a8bf15f47d3b12a2cc71e1798e32691d7424" exitCode=255 Mar 18 00:08:19 crc kubenswrapper[4983]: I0318 00:08:19.993745 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"772ad4d5c7862487c75cca409c63a8bf15f47d3b12a2cc71e1798e32691d7424"} Mar 18 00:08:19 crc kubenswrapper[4983]: I0318 00:08:19.993792 4983 scope.go:117] "RemoveContainer" containerID="350b6156522c018873b37e48f03b0b90eae9d690480d8f4bcf44e1c751443bae" Mar 18 00:08:19 crc kubenswrapper[4983]: I0318 00:08:19.994096 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:19 crc kubenswrapper[4983]: I0318 00:08:19.995752 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:19 crc kubenswrapper[4983]: I0318 00:08:19.995808 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:19 crc kubenswrapper[4983]: I0318 00:08:19.995827 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:19 crc kubenswrapper[4983]: I0318 00:08:19.997297 4983 scope.go:117] "RemoveContainer" containerID="772ad4d5c7862487c75cca409c63a8bf15f47d3b12a2cc71e1798e32691d7424" Mar 18 00:08:19 crc kubenswrapper[4983]: E0318 00:08:19.997623 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:08:20 crc kubenswrapper[4983]: W0318 00:08:20.185304 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:20Z is after 2026-02-23T05:33:13Z Mar 18 00:08:20 crc kubenswrapper[4983]: E0318 00:08:20.185409 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:20Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:20 crc kubenswrapper[4983]: I0318 00:08:20.758689 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:20Z is after 2026-02-23T05:33:13Z Mar 18 00:08:20 crc kubenswrapper[4983]: I0318 00:08:20.820526 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:20 crc kubenswrapper[4983]: I0318 00:08:20.999118 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 18 00:08:21 crc kubenswrapper[4983]: I0318 00:08:21.002897 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:21 crc kubenswrapper[4983]: I0318 00:08:21.004126 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:21 crc kubenswrapper[4983]: I0318 00:08:21.004421 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:21 crc kubenswrapper[4983]: I0318 00:08:21.004757 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:21 crc kubenswrapper[4983]: I0318 00:08:21.005748 4983 scope.go:117] "RemoveContainer" containerID="772ad4d5c7862487c75cca409c63a8bf15f47d3b12a2cc71e1798e32691d7424" Mar 18 00:08:21 crc kubenswrapper[4983]: E0318 00:08:21.006216 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:08:21 crc kubenswrapper[4983]: I0318 00:08:21.090440 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:21 crc kubenswrapper[4983]: W0318 00:08:21.496087 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:21Z is after 2026-02-23T05:33:13Z Mar 18 00:08:21 crc kubenswrapper[4983]: E0318 00:08:21.496260 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:21Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:21 crc kubenswrapper[4983]: I0318 00:08:21.757205 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:21Z is after 2026-02-23T05:33:13Z Mar 18 00:08:22 crc kubenswrapper[4983]: I0318 00:08:22.006463 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:22 crc kubenswrapper[4983]: I0318 00:08:22.011810 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:22 crc kubenswrapper[4983]: I0318 00:08:22.011900 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:22 crc kubenswrapper[4983]: I0318 00:08:22.011958 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:22 crc kubenswrapper[4983]: I0318 00:08:22.014073 4983 scope.go:117] "RemoveContainer" containerID="772ad4d5c7862487c75cca409c63a8bf15f47d3b12a2cc71e1798e32691d7424" Mar 18 00:08:22 crc kubenswrapper[4983]: E0318 00:08:22.014449 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:08:22 crc kubenswrapper[4983]: I0318 00:08:22.016101 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:22 crc kubenswrapper[4983]: W0318 00:08:22.191297 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:22Z is after 2026-02-23T05:33:13Z Mar 18 00:08:22 crc kubenswrapper[4983]: E0318 00:08:22.191522 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:22Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:22 crc kubenswrapper[4983]: I0318 00:08:22.204969 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 18 00:08:22 crc kubenswrapper[4983]: I0318 00:08:22.205286 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:22 crc kubenswrapper[4983]: I0318 00:08:22.207862 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:22 crc kubenswrapper[4983]: I0318 00:08:22.207968 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:22 crc kubenswrapper[4983]: I0318 00:08:22.207995 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:22 crc kubenswrapper[4983]: I0318 00:08:22.237318 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 18 00:08:22 crc kubenswrapper[4983]: I0318 00:08:22.757970 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:22Z is after 2026-02-23T05:33:13Z Mar 18 00:08:22 crc kubenswrapper[4983]: E0318 00:08:22.927525 4983 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.008785 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.008794 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.011044 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.011141 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.011166 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.011508 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.011555 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.011571 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.012160 4983 scope.go:117] "RemoveContainer" containerID="772ad4d5c7862487c75cca409c63a8bf15f47d3b12a2cc71e1798e32691d7424" Mar 18 00:08:23 crc kubenswrapper[4983]: E0318 00:08:23.012473 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.674570 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.676425 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.676494 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.676517 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.676558 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:08:23 crc kubenswrapper[4983]: E0318 00:08:23.676886 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:23Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 18 00:08:23 crc kubenswrapper[4983]: E0318 00:08:23.681380 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:23Z is after 2026-02-23T05:33:13Z" node="crc" Mar 18 00:08:23 crc kubenswrapper[4983]: I0318 00:08:23.757965 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:23Z is after 2026-02-23T05:33:13Z Mar 18 00:08:24 crc kubenswrapper[4983]: I0318 00:08:24.758833 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:24Z is after 2026-02-23T05:33:13Z Mar 18 00:08:25 crc kubenswrapper[4983]: I0318 00:08:25.760427 4983 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 00:08:25 crc kubenswrapper[4983]: I0318 00:08:25.764216 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:25Z is after 2026-02-23T05:33:13Z Mar 18 00:08:25 crc kubenswrapper[4983]: E0318 00:08:25.766187 4983 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:25Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:26 crc kubenswrapper[4983]: I0318 00:08:26.759730 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:26Z is after 2026-02-23T05:33:13Z Mar 18 00:08:27 crc kubenswrapper[4983]: E0318 00:08:27.273205 4983 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:27Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189dc6dd778a15d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.749453777 +0000 UTC m=+0.646828417,LastTimestamp:2026-03-18 00:08:02.749453777 +0000 UTC m=+0.646828417,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:27 crc kubenswrapper[4983]: I0318 00:08:27.363758 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:27 crc kubenswrapper[4983]: I0318 00:08:27.364368 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:27 crc kubenswrapper[4983]: I0318 00:08:27.366291 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:27 crc kubenswrapper[4983]: I0318 00:08:27.366343 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:27 crc kubenswrapper[4983]: I0318 00:08:27.366366 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:27 crc kubenswrapper[4983]: I0318 00:08:27.367220 4983 scope.go:117] "RemoveContainer" containerID="772ad4d5c7862487c75cca409c63a8bf15f47d3b12a2cc71e1798e32691d7424" Mar 18 00:08:27 crc kubenswrapper[4983]: E0318 00:08:27.367521 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:08:27 crc kubenswrapper[4983]: W0318 00:08:27.462043 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:27Z is after 2026-02-23T05:33:13Z Mar 18 00:08:27 crc kubenswrapper[4983]: E0318 00:08:27.462143 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:27Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:27 crc kubenswrapper[4983]: I0318 00:08:27.757719 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:27Z is after 2026-02-23T05:33:13Z Mar 18 00:08:28 crc kubenswrapper[4983]: I0318 00:08:28.350849 4983 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 00:08:28 crc kubenswrapper[4983]: I0318 00:08:28.350924 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 00:08:28 crc kubenswrapper[4983]: I0318 00:08:28.758369 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:28Z is after 2026-02-23T05:33:13Z Mar 18 00:08:29 crc kubenswrapper[4983]: W0318 00:08:29.166059 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:29Z is after 2026-02-23T05:33:13Z Mar 18 00:08:29 crc kubenswrapper[4983]: E0318 00:08:29.166158 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:29Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:29 crc kubenswrapper[4983]: W0318 00:08:29.629798 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:29Z is after 2026-02-23T05:33:13Z Mar 18 00:08:29 crc kubenswrapper[4983]: E0318 00:08:29.630664 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:29Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:29 crc kubenswrapper[4983]: I0318 00:08:29.758502 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:29Z is after 2026-02-23T05:33:13Z Mar 18 00:08:30 crc kubenswrapper[4983]: I0318 00:08:30.681856 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:30 crc kubenswrapper[4983]: I0318 00:08:30.683434 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:30 crc kubenswrapper[4983]: I0318 00:08:30.683483 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:30 crc kubenswrapper[4983]: I0318 00:08:30.683500 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:30 crc kubenswrapper[4983]: I0318 00:08:30.683534 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:08:30 crc kubenswrapper[4983]: E0318 00:08:30.683982 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:30Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 18 00:08:30 crc kubenswrapper[4983]: E0318 00:08:30.688342 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:30Z is after 2026-02-23T05:33:13Z" node="crc" Mar 18 00:08:30 crc kubenswrapper[4983]: I0318 00:08:30.758126 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:30Z is after 2026-02-23T05:33:13Z Mar 18 00:08:31 crc kubenswrapper[4983]: I0318 00:08:31.758611 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:31Z is after 2026-02-23T05:33:13Z Mar 18 00:08:32 crc kubenswrapper[4983]: I0318 00:08:32.758482 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:32Z is after 2026-02-23T05:33:13Z Mar 18 00:08:32 crc kubenswrapper[4983]: E0318 00:08:32.927730 4983 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 00:08:33 crc kubenswrapper[4983]: W0318 00:08:33.511374 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:33Z is after 2026-02-23T05:33:13Z Mar 18 00:08:33 crc kubenswrapper[4983]: E0318 00:08:33.511487 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:33Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:33 crc kubenswrapper[4983]: I0318 00:08:33.759323 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:33Z is after 2026-02-23T05:33:13Z Mar 18 00:08:34 crc kubenswrapper[4983]: I0318 00:08:34.758587 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z Mar 18 00:08:35 crc kubenswrapper[4983]: I0318 00:08:35.412176 4983 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:48610->192.168.126.11:10357: read: connection reset by peer" start-of-body= Mar 18 00:08:35 crc kubenswrapper[4983]: I0318 00:08:35.412254 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:48610->192.168.126.11:10357: read: connection reset by peer" Mar 18 00:08:35 crc kubenswrapper[4983]: I0318 00:08:35.412331 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:35 crc kubenswrapper[4983]: I0318 00:08:35.412503 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:35 crc kubenswrapper[4983]: I0318 00:08:35.414377 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:35 crc kubenswrapper[4983]: I0318 00:08:35.414522 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:35 crc kubenswrapper[4983]: I0318 00:08:35.414549 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:35 crc kubenswrapper[4983]: I0318 00:08:35.415326 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 18 00:08:35 crc kubenswrapper[4983]: I0318 00:08:35.415574 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d" gracePeriod=30 Mar 18 00:08:35 crc kubenswrapper[4983]: I0318 00:08:35.759966 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:35Z is after 2026-02-23T05:33:13Z Mar 18 00:08:36 crc kubenswrapper[4983]: I0318 00:08:36.051903 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 18 00:08:36 crc kubenswrapper[4983]: I0318 00:08:36.053025 4983 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d" exitCode=255 Mar 18 00:08:36 crc kubenswrapper[4983]: I0318 00:08:36.053085 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d"} Mar 18 00:08:36 crc kubenswrapper[4983]: I0318 00:08:36.053144 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab"} Mar 18 00:08:36 crc kubenswrapper[4983]: I0318 00:08:36.053293 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:36 crc kubenswrapper[4983]: I0318 00:08:36.054569 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:36 crc kubenswrapper[4983]: I0318 00:08:36.054634 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:36 crc kubenswrapper[4983]: I0318 00:08:36.054659 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:36 crc kubenswrapper[4983]: I0318 00:08:36.079535 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:36 crc kubenswrapper[4983]: I0318 00:08:36.757697 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:36Z is after 2026-02-23T05:33:13Z Mar 18 00:08:37 crc kubenswrapper[4983]: I0318 00:08:37.056043 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:37 crc kubenswrapper[4983]: I0318 00:08:37.057424 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:37 crc kubenswrapper[4983]: I0318 00:08:37.057476 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:37 crc kubenswrapper[4983]: I0318 00:08:37.057494 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:37 crc kubenswrapper[4983]: E0318 00:08:37.280374 4983 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:37Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189dc6dd778a15d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.749453777 +0000 UTC m=+0.646828417,LastTimestamp:2026-03-18 00:08:02.749453777 +0000 UTC m=+0.646828417,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:37 crc kubenswrapper[4983]: I0318 00:08:37.688609 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:37 crc kubenswrapper[4983]: I0318 00:08:37.690547 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:37 crc kubenswrapper[4983]: I0318 00:08:37.690658 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:37 crc kubenswrapper[4983]: I0318 00:08:37.690682 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:37 crc kubenswrapper[4983]: I0318 00:08:37.690734 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:08:37 crc kubenswrapper[4983]: E0318 00:08:37.691711 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:37Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 18 00:08:37 crc kubenswrapper[4983]: E0318 00:08:37.696219 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:37Z is after 2026-02-23T05:33:13Z" node="crc" Mar 18 00:08:37 crc kubenswrapper[4983]: I0318 00:08:37.758750 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:37Z is after 2026-02-23T05:33:13Z Mar 18 00:08:38 crc kubenswrapper[4983]: I0318 00:08:38.757723 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:38Z is after 2026-02-23T05:33:13Z Mar 18 00:08:39 crc kubenswrapper[4983]: I0318 00:08:39.758547 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:39Z is after 2026-02-23T05:33:13Z Mar 18 00:08:39 crc kubenswrapper[4983]: I0318 00:08:39.843672 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:39 crc kubenswrapper[4983]: I0318 00:08:39.846561 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:39 crc kubenswrapper[4983]: I0318 00:08:39.846997 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:39 crc kubenswrapper[4983]: I0318 00:08:39.847207 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:39 crc kubenswrapper[4983]: I0318 00:08:39.848452 4983 scope.go:117] "RemoveContainer" containerID="772ad4d5c7862487c75cca409c63a8bf15f47d3b12a2cc71e1798e32691d7424" Mar 18 00:08:40 crc kubenswrapper[4983]: I0318 00:08:40.758878 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:40Z is after 2026-02-23T05:33:13Z Mar 18 00:08:41 crc kubenswrapper[4983]: I0318 00:08:41.069228 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 18 00:08:41 crc kubenswrapper[4983]: I0318 00:08:41.070352 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 18 00:08:41 crc kubenswrapper[4983]: I0318 00:08:41.073782 4983 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3ae69de2b7a7acff7586d0220e4f2487998f8552d7a6b1fb82efe01a56c9b279" exitCode=255 Mar 18 00:08:41 crc kubenswrapper[4983]: I0318 00:08:41.073897 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3ae69de2b7a7acff7586d0220e4f2487998f8552d7a6b1fb82efe01a56c9b279"} Mar 18 00:08:41 crc kubenswrapper[4983]: I0318 00:08:41.074041 4983 scope.go:117] "RemoveContainer" containerID="772ad4d5c7862487c75cca409c63a8bf15f47d3b12a2cc71e1798e32691d7424" Mar 18 00:08:41 crc kubenswrapper[4983]: I0318 00:08:41.074288 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:41 crc kubenswrapper[4983]: I0318 00:08:41.075810 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:41 crc kubenswrapper[4983]: I0318 00:08:41.075870 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:41 crc kubenswrapper[4983]: I0318 00:08:41.075890 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:41 crc kubenswrapper[4983]: I0318 00:08:41.076849 4983 scope.go:117] "RemoveContainer" containerID="3ae69de2b7a7acff7586d0220e4f2487998f8552d7a6b1fb82efe01a56c9b279" Mar 18 00:08:41 crc kubenswrapper[4983]: E0318 00:08:41.077177 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:08:41 crc kubenswrapper[4983]: I0318 00:08:41.758242 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:41Z is after 2026-02-23T05:33:13Z Mar 18 00:08:42 crc kubenswrapper[4983]: I0318 00:08:42.079348 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 18 00:08:42 crc kubenswrapper[4983]: I0318 00:08:42.703489 4983 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 00:08:42 crc kubenswrapper[4983]: E0318 00:08:42.709563 4983 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:42Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:42 crc kubenswrapper[4983]: E0318 00:08:42.710864 4983 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 18 00:08:42 crc kubenswrapper[4983]: I0318 00:08:42.759215 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:42Z is after 2026-02-23T05:33:13Z Mar 18 00:08:42 crc kubenswrapper[4983]: E0318 00:08:42.928677 4983 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 00:08:43 crc kubenswrapper[4983]: I0318 00:08:43.759215 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:43Z is after 2026-02-23T05:33:13Z Mar 18 00:08:44 crc kubenswrapper[4983]: I0318 00:08:44.697145 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:44 crc kubenswrapper[4983]: E0318 00:08:44.697752 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:44Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 18 00:08:44 crc kubenswrapper[4983]: I0318 00:08:44.699398 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:44 crc kubenswrapper[4983]: I0318 00:08:44.699463 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:44 crc kubenswrapper[4983]: I0318 00:08:44.699502 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:44 crc kubenswrapper[4983]: I0318 00:08:44.699572 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:08:44 crc kubenswrapper[4983]: E0318 00:08:44.706344 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:44Z is after 2026-02-23T05:33:13Z" node="crc" Mar 18 00:08:44 crc kubenswrapper[4983]: I0318 00:08:44.759035 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:44Z is after 2026-02-23T05:33:13Z Mar 18 00:08:44 crc kubenswrapper[4983]: W0318 00:08:44.928153 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:44Z is after 2026-02-23T05:33:13Z Mar 18 00:08:44 crc kubenswrapper[4983]: E0318 00:08:44.928258 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:44Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:45 crc kubenswrapper[4983]: I0318 00:08:45.350064 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:08:45 crc kubenswrapper[4983]: I0318 00:08:45.350335 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:45 crc kubenswrapper[4983]: I0318 00:08:45.352059 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:45 crc kubenswrapper[4983]: I0318 00:08:45.352194 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:45 crc kubenswrapper[4983]: I0318 00:08:45.352219 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:45 crc kubenswrapper[4983]: W0318 00:08:45.754477 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:45Z is after 2026-02-23T05:33:13Z Mar 18 00:08:45 crc kubenswrapper[4983]: E0318 00:08:45.754597 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:45Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:45 crc kubenswrapper[4983]: I0318 00:08:45.757888 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:45Z is after 2026-02-23T05:33:13Z Mar 18 00:08:46 crc kubenswrapper[4983]: I0318 00:08:46.756614 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:46Z is after 2026-02-23T05:33:13Z Mar 18 00:08:47 crc kubenswrapper[4983]: E0318 00:08:47.286388 4983 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:47Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189dc6dd778a15d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.749453777 +0000 UTC m=+0.646828417,LastTimestamp:2026-03-18 00:08:02.749453777 +0000 UTC m=+0.646828417,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:47 crc kubenswrapper[4983]: I0318 00:08:47.364008 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:47 crc kubenswrapper[4983]: I0318 00:08:47.364269 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:47 crc kubenswrapper[4983]: I0318 00:08:47.366103 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:47 crc kubenswrapper[4983]: I0318 00:08:47.366173 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:47 crc kubenswrapper[4983]: I0318 00:08:47.366201 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:47 crc kubenswrapper[4983]: I0318 00:08:47.367187 4983 scope.go:117] "RemoveContainer" containerID="3ae69de2b7a7acff7586d0220e4f2487998f8552d7a6b1fb82efe01a56c9b279" Mar 18 00:08:47 crc kubenswrapper[4983]: E0318 00:08:47.367479 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:08:47 crc kubenswrapper[4983]: I0318 00:08:47.757559 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:47Z is after 2026-02-23T05:33:13Z Mar 18 00:08:48 crc kubenswrapper[4983]: I0318 00:08:48.350655 4983 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 00:08:48 crc kubenswrapper[4983]: I0318 00:08:48.351113 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 00:08:48 crc kubenswrapper[4983]: I0318 00:08:48.759284 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:48Z is after 2026-02-23T05:33:13Z Mar 18 00:08:49 crc kubenswrapper[4983]: W0318 00:08:49.156348 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:49Z is after 2026-02-23T05:33:13Z Mar 18 00:08:49 crc kubenswrapper[4983]: E0318 00:08:49.156456 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:49Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 00:08:49 crc kubenswrapper[4983]: I0318 00:08:49.757876 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:49Z is after 2026-02-23T05:33:13Z Mar 18 00:08:50 crc kubenswrapper[4983]: I0318 00:08:50.758447 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:50Z is after 2026-02-23T05:33:13Z Mar 18 00:08:50 crc kubenswrapper[4983]: I0318 00:08:50.820823 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:08:50 crc kubenswrapper[4983]: I0318 00:08:50.821075 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:50 crc kubenswrapper[4983]: I0318 00:08:50.822566 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:50 crc kubenswrapper[4983]: I0318 00:08:50.822613 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:50 crc kubenswrapper[4983]: I0318 00:08:50.822635 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:50 crc kubenswrapper[4983]: I0318 00:08:50.823398 4983 scope.go:117] "RemoveContainer" containerID="3ae69de2b7a7acff7586d0220e4f2487998f8552d7a6b1fb82efe01a56c9b279" Mar 18 00:08:50 crc kubenswrapper[4983]: E0318 00:08:50.823703 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:08:51 crc kubenswrapper[4983]: E0318 00:08:51.703143 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:51Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 18 00:08:51 crc kubenswrapper[4983]: I0318 00:08:51.707356 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:51 crc kubenswrapper[4983]: I0318 00:08:51.708311 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:51 crc kubenswrapper[4983]: I0318 00:08:51.708333 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:51 crc kubenswrapper[4983]: I0318 00:08:51.708342 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:51 crc kubenswrapper[4983]: I0318 00:08:51.708363 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:08:51 crc kubenswrapper[4983]: E0318 00:08:51.711311 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:51Z is after 2026-02-23T05:33:13Z" node="crc" Mar 18 00:08:51 crc kubenswrapper[4983]: I0318 00:08:51.758318 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:51Z is after 2026-02-23T05:33:13Z Mar 18 00:08:52 crc kubenswrapper[4983]: I0318 00:08:52.761552 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:52Z is after 2026-02-23T05:33:13Z Mar 18 00:08:52 crc kubenswrapper[4983]: E0318 00:08:52.928947 4983 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 00:08:53 crc kubenswrapper[4983]: I0318 00:08:53.761443 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:08:54 crc kubenswrapper[4983]: W0318 00:08:54.578631 4983 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 18 00:08:54 crc kubenswrapper[4983]: E0318 00:08:54.578743 4983 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 18 00:08:54 crc kubenswrapper[4983]: I0318 00:08:54.759338 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:08:55 crc kubenswrapper[4983]: I0318 00:08:55.760580 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:08:56 crc kubenswrapper[4983]: I0318 00:08:56.760307 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.294570 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd778a15d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.749453777 +0000 UTC m=+0.646828417,LastTimestamp:2026-03-18 00:08:02.749453777 +0000 UTC m=+0.646828417,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.301714 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a0bf1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832288753 +0000 UTC m=+0.729663393,LastTimestamp:2026-03-18 00:08:02.832288753 +0000 UTC m=+0.729663393,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.307494 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a69a8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832312744 +0000 UTC m=+0.729687384,LastTimestamp:2026-03-18 00:08:02.832312744 +0000 UTC m=+0.729687384,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.313726 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7aa6c0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832328384 +0000 UTC m=+0.729703034,LastTimestamp:2026-03-18 00:08:02.832328384 +0000 UTC m=+0.729703034,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.318648 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd8171629f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.915607199 +0000 UTC m=+0.812981849,LastTimestamp:2026-03-18 00:08:02.915607199 +0000 UTC m=+0.812981849,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.324421 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a0bf1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a0bf1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832288753 +0000 UTC m=+0.729663393,LastTimestamp:2026-03-18 00:08:02.945844676 +0000 UTC m=+0.843219326,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.329301 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a69a8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a69a8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832312744 +0000 UTC m=+0.729687384,LastTimestamp:2026-03-18 00:08:02.945882317 +0000 UTC m=+0.843256967,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.333832 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7aa6c0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7aa6c0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832328384 +0000 UTC m=+0.729703034,LastTimestamp:2026-03-18 00:08:02.945901027 +0000 UTC m=+0.843275677,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.338883 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a0bf1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a0bf1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832288753 +0000 UTC m=+0.729663393,LastTimestamp:2026-03-18 00:08:02.947514427 +0000 UTC m=+0.844889047,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.346048 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a69a8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a69a8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832312744 +0000 UTC m=+0.729687384,LastTimestamp:2026-03-18 00:08:02.947533017 +0000 UTC m=+0.844907637,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.352640 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7aa6c0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7aa6c0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832328384 +0000 UTC m=+0.729703034,LastTimestamp:2026-03-18 00:08:02.947547348 +0000 UTC m=+0.844921968,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.358485 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a0bf1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a0bf1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832288753 +0000 UTC m=+0.729663393,LastTimestamp:2026-03-18 00:08:02.949415852 +0000 UTC m=+0.846790472,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.363752 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a69a8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a69a8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832312744 +0000 UTC m=+0.729687384,LastTimestamp:2026-03-18 00:08:02.949460993 +0000 UTC m=+0.846835613,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.370494 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7aa6c0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7aa6c0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832328384 +0000 UTC m=+0.729703034,LastTimestamp:2026-03-18 00:08:02.949476453 +0000 UTC m=+0.846851083,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.376386 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a0bf1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a0bf1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832288753 +0000 UTC m=+0.729663393,LastTimestamp:2026-03-18 00:08:02.950184816 +0000 UTC m=+0.847559436,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.381652 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a69a8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a69a8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832312744 +0000 UTC m=+0.729687384,LastTimestamp:2026-03-18 00:08:02.950210047 +0000 UTC m=+0.847584667,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.387558 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7aa6c0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7aa6c0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832328384 +0000 UTC m=+0.729703034,LastTimestamp:2026-03-18 00:08:02.950223907 +0000 UTC m=+0.847598527,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.393298 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a0bf1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a0bf1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832288753 +0000 UTC m=+0.729663393,LastTimestamp:2026-03-18 00:08:02.95145988 +0000 UTC m=+0.848834520,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.397766 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a69a8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a69a8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832312744 +0000 UTC m=+0.729687384,LastTimestamp:2026-03-18 00:08:02.95149058 +0000 UTC m=+0.848865220,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.403969 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7aa6c0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7aa6c0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832328384 +0000 UTC m=+0.729703034,LastTimestamp:2026-03-18 00:08:02.951509831 +0000 UTC m=+0.848884481,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.407495 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a0bf1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a0bf1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832288753 +0000 UTC m=+0.729663393,LastTimestamp:2026-03-18 00:08:02.951593602 +0000 UTC m=+0.848968222,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.410576 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a69a8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a69a8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832312744 +0000 UTC m=+0.729687384,LastTimestamp:2026-03-18 00:08:02.951608452 +0000 UTC m=+0.848983072,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.412438 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7aa6c0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7aa6c0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832328384 +0000 UTC m=+0.729703034,LastTimestamp:2026-03-18 00:08:02.951619193 +0000 UTC m=+0.848993813,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.416850 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a0bf1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a0bf1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832288753 +0000 UTC m=+0.729663393,LastTimestamp:2026-03-18 00:08:02.952276855 +0000 UTC m=+0.849651505,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.419133 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189dc6dd7c7a69a8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189dc6dd7c7a69a8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:02.832312744 +0000 UTC m=+0.729687384,LastTimestamp:2026-03-18 00:08:02.952315505 +0000 UTC m=+0.849690155,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.426063 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6dd9bbb5771 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.356661617 +0000 UTC m=+1.254036257,LastTimestamp:2026-03-18 00:08:03.356661617 +0000 UTC m=+1.254036257,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.431036 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6dd9bcbd21d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.357741597 +0000 UTC m=+1.255116247,LastTimestamp:2026-03-18 00:08:03.357741597 +0000 UTC m=+1.255116247,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.435699 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6dd9c124202 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.362357762 +0000 UTC m=+1.259732412,LastTimestamp:2026-03-18 00:08:03.362357762 +0000 UTC m=+1.259732412,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.441203 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189dc6dd9ce84364 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.37638282 +0000 UTC m=+1.273757460,LastTimestamp:2026-03-18 00:08:03.37638282 +0000 UTC m=+1.273757460,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.447439 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189dc6dd9f096de3 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.412110819 +0000 UTC m=+1.309485469,LastTimestamp:2026-03-18 00:08:03.412110819 +0000 UTC m=+1.309485469,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.454191 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6ddbfd9faa8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.962649256 +0000 UTC m=+1.860023916,LastTimestamp:2026-03-18 00:08:03.962649256 +0000 UTC m=+1.860023916,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.460349 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189dc6ddbfdda9d6 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.96289071 +0000 UTC m=+1.860265360,LastTimestamp:2026-03-18 00:08:03.96289071 +0000 UTC m=+1.860265360,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.466218 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6ddbfe5191d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.963377949 +0000 UTC m=+1.860752599,LastTimestamp:2026-03-18 00:08:03.963377949 +0000 UTC m=+1.860752599,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.472356 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189dc6ddc0721e2c openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.97261982 +0000 UTC m=+1.869994460,LastTimestamp:2026-03-18 00:08:03.97261982 +0000 UTC m=+1.869994460,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.478667 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189dc6ddc08034ff openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.973543167 +0000 UTC m=+1.870917817,LastTimestamp:2026-03-18 00:08:03.973543167 +0000 UTC m=+1.870917817,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.483347 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6ddc084e37e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.973849982 +0000 UTC m=+1.871224622,LastTimestamp:2026-03-18 00:08:03.973849982 +0000 UTC m=+1.871224622,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.487554 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6ddc105c13e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.982295358 +0000 UTC m=+1.879669998,LastTimestamp:2026-03-18 00:08:03.982295358 +0000 UTC m=+1.879669998,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.492697 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6ddc12b7f42 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.984768834 +0000 UTC m=+1.882143484,LastTimestamp:2026-03-18 00:08:03.984768834 +0000 UTC m=+1.882143484,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.497793 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6ddc15a0cd2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.98781973 +0000 UTC m=+1.885194350,LastTimestamp:2026-03-18 00:08:03.98781973 +0000 UTC m=+1.885194350,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.503422 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6ddc1af7ba7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.993418663 +0000 UTC m=+1.890793323,LastTimestamp:2026-03-18 00:08:03.993418663 +0000 UTC m=+1.890793323,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.512068 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189dc6ddc1b67e42 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.993878082 +0000 UTC m=+1.891252702,LastTimestamp:2026-03-18 00:08:03.993878082 +0000 UTC m=+1.891252702,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.516774 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6ddd5493528 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.322260264 +0000 UTC m=+2.219634904,LastTimestamp:2026-03-18 00:08:04.322260264 +0000 UTC m=+2.219634904,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.522962 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6ddd61f6129 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.336296233 +0000 UTC m=+2.233670883,LastTimestamp:2026-03-18 00:08:04.336296233 +0000 UTC m=+2.233670883,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.528567 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6ddd634e955 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.337707349 +0000 UTC m=+2.235081999,LastTimestamp:2026-03-18 00:08:04.337707349 +0000 UTC m=+2.235081999,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.534867 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6dde6113af3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.603804403 +0000 UTC m=+2.501179023,LastTimestamp:2026-03-18 00:08:04.603804403 +0000 UTC m=+2.501179023,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.540667 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6dde6cca5dd openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.616087005 +0000 UTC m=+2.513461655,LastTimestamp:2026-03-18 00:08:04.616087005 +0000 UTC m=+2.513461655,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.545184 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6dde6e6042c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.617749548 +0000 UTC m=+2.515124208,LastTimestamp:2026-03-18 00:08:04.617749548 +0000 UTC m=+2.515124208,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.550526 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6ddf4de0bf4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.852108276 +0000 UTC m=+2.749482886,LastTimestamp:2026-03-18 00:08:04.852108276 +0000 UTC m=+2.749482886,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.556089 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6ddf5a35feb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.865040363 +0000 UTC m=+2.762415003,LastTimestamp:2026-03-18 00:08:04.865040363 +0000 UTC m=+2.762415003,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.563958 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6ddf5ce7abc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.867865276 +0000 UTC m=+2.765239896,LastTimestamp:2026-03-18 00:08:04.867865276 +0000 UTC m=+2.765239896,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.568720 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6ddf5ec0e02 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.869803522 +0000 UTC m=+2.767178132,LastTimestamp:2026-03-18 00:08:04.869803522 +0000 UTC m=+2.767178132,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.573736 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189dc6ddf64258d1 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.875458769 +0000 UTC m=+2.772833379,LastTimestamp:2026-03-18 00:08:04.875458769 +0000 UTC m=+2.772833379,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.581546 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189dc6ddf6691a62 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.87799869 +0000 UTC m=+2.775373300,LastTimestamp:2026-03-18 00:08:04.87799869 +0000 UTC m=+2.775373300,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.587430 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189dc6de046ba674 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.113046644 +0000 UTC m=+3.010421254,LastTimestamp:2026-03-18 00:08:05.113046644 +0000 UTC m=+3.010421254,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.593818 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6de0470ca98 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.113383576 +0000 UTC m=+3.010758186,LastTimestamp:2026-03-18 00:08:05.113383576 +0000 UTC m=+3.010758186,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.600618 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de04772a22 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.11380125 +0000 UTC m=+3.011175860,LastTimestamp:2026-03-18 00:08:05.11380125 +0000 UTC m=+3.011175860,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.607579 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189dc6de047ab036 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.114032182 +0000 UTC m=+3.011406792,LastTimestamp:2026-03-18 00:08:05.114032182 +0000 UTC m=+3.011406792,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.613182 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de051d6b4e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.12469691 +0000 UTC m=+3.022071520,LastTimestamp:2026-03-18 00:08:05.12469691 +0000 UTC m=+3.022071520,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.618987 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de052e53ef openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.125805039 +0000 UTC m=+3.023179649,LastTimestamp:2026-03-18 00:08:05.125805039 +0000 UTC m=+3.023179649,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.622009 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189dc6de0540ec67 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.127023719 +0000 UTC m=+3.024398329,LastTimestamp:2026-03-18 00:08:05.127023719 +0000 UTC m=+3.024398329,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.623488 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189dc6de055415ae openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.12827947 +0000 UTC m=+3.025654070,LastTimestamp:2026-03-18 00:08:05.12827947 +0000 UTC m=+3.025654070,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.630356 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189dc6de05931028 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.132406824 +0000 UTC m=+3.029781434,LastTimestamp:2026-03-18 00:08:05.132406824 +0000 UTC m=+3.029781434,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.635103 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6de06c705f5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.152589301 +0000 UTC m=+3.049963911,LastTimestamp:2026-03-18 00:08:05.152589301 +0000 UTC m=+3.049963911,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.639203 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189dc6de0fc4d116 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.303439638 +0000 UTC m=+3.200814248,LastTimestamp:2026-03-18 00:08:05.303439638 +0000 UTC m=+3.200814248,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.645507 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de104aea6e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.31222795 +0000 UTC m=+3.209602550,LastTimestamp:2026-03-18 00:08:05.31222795 +0000 UTC m=+3.209602550,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.650848 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189dc6de10d629e0 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.321353696 +0000 UTC m=+3.218728306,LastTimestamp:2026-03-18 00:08:05.321353696 +0000 UTC m=+3.218728306,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.657699 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189dc6de10ef02bf openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.322982079 +0000 UTC m=+3.220356689,LastTimestamp:2026-03-18 00:08:05.322982079 +0000 UTC m=+3.220356689,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.661841 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de10f63be7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.323455463 +0000 UTC m=+3.220830073,LastTimestamp:2026-03-18 00:08:05.323455463 +0000 UTC m=+3.220830073,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.666760 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de1104248f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.324366991 +0000 UTC m=+3.221741611,LastTimestamp:2026-03-18 00:08:05.324366991 +0000 UTC m=+3.221741611,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.671471 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de1b05f53d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.492258109 +0000 UTC m=+3.389632719,LastTimestamp:2026-03-18 00:08:05.492258109 +0000 UTC m=+3.389632719,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.676657 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de1c2044a8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.510759592 +0000 UTC m=+3.408134202,LastTimestamp:2026-03-18 00:08:05.510759592 +0000 UTC m=+3.408134202,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.681719 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de1c2f8a4c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.51176046 +0000 UTC m=+3.409135070,LastTimestamp:2026-03-18 00:08:05.51176046 +0000 UTC m=+3.409135070,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.686604 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189dc6de1eefc56e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.557912942 +0000 UTC m=+3.455287552,LastTimestamp:2026-03-18 00:08:05.557912942 +0000 UTC m=+3.455287552,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.691907 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189dc6de1fa12904 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.569538308 +0000 UTC m=+3.466912938,LastTimestamp:2026-03-18 00:08:05.569538308 +0000 UTC m=+3.466912938,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.696738 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de28103f80 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.711036288 +0000 UTC m=+3.608410888,LastTimestamp:2026-03-18 00:08:05.711036288 +0000 UTC m=+3.608410888,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.701403 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de291ab3be openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.728498622 +0000 UTC m=+3.625873272,LastTimestamp:2026-03-18 00:08:05.728498622 +0000 UTC m=+3.625873272,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.706173 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de292c3b16 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.729647382 +0000 UTC m=+3.627021992,LastTimestamp:2026-03-18 00:08:05.729647382 +0000 UTC m=+3.627021992,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.712967 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6de343be208 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.915222536 +0000 UTC m=+3.812597166,LastTimestamp:2026-03-18 00:08:05.915222536 +0000 UTC m=+3.812597166,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.718094 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de38a9924e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.98951995 +0000 UTC m=+3.886894570,LastTimestamp:2026-03-18 00:08:05.98951995 +0000 UTC m=+3.886894570,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.723174 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de39f1dd94 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:06.011035028 +0000 UTC m=+3.908409648,LastTimestamp:2026-03-18 00:08:06.011035028 +0000 UTC m=+3.908409648,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.730062 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6de401d49ac openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:06.114544044 +0000 UTC m=+4.011918664,LastTimestamp:2026-03-18 00:08:06.114544044 +0000 UTC m=+4.011918664,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.735196 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6de4114de28 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:06.130769448 +0000 UTC m=+4.028144068,LastTimestamp:2026-03-18 00:08:06.130769448 +0000 UTC m=+4.028144068,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.741484 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6de7107c2ee openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:06.935216878 +0000 UTC m=+4.832591518,LastTimestamp:2026-03-18 00:08:06.935216878 +0000 UTC m=+4.832591518,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.748315 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6de7faae4d4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:07.180788948 +0000 UTC m=+5.078163588,LastTimestamp:2026-03-18 00:08:07.180788948 +0000 UTC m=+5.078163588,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.757712 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6de806c0c81 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:07.193447553 +0000 UTC m=+5.090822203,LastTimestamp:2026-03-18 00:08:07.193447553 +0000 UTC m=+5.090822203,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: I0318 00:08:57.758088 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.767003 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6de8080a6d8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:07.194797784 +0000 UTC m=+5.092172424,LastTimestamp:2026-03-18 00:08:07.194797784 +0000 UTC m=+5.092172424,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.776406 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6de8f09a836 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:07.43865759 +0000 UTC m=+5.336032240,LastTimestamp:2026-03-18 00:08:07.43865759 +0000 UTC m=+5.336032240,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.783328 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6de8ffde5b1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:07.454664113 +0000 UTC m=+5.352038763,LastTimestamp:2026-03-18 00:08:07.454664113 +0000 UTC m=+5.352038763,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.789445 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6de90171f8f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:07.456317327 +0000 UTC m=+5.353691967,LastTimestamp:2026-03-18 00:08:07.456317327 +0000 UTC m=+5.353691967,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.794800 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6dea06498a0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:07.729830048 +0000 UTC m=+5.627204698,LastTimestamp:2026-03-18 00:08:07.729830048 +0000 UTC m=+5.627204698,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.801099 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6dea1596364 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:07.74587274 +0000 UTC m=+5.643247390,LastTimestamp:2026-03-18 00:08:07.74587274 +0000 UTC m=+5.643247390,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.807048 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6dea1748a1d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:07.747652125 +0000 UTC m=+5.645026775,LastTimestamp:2026-03-18 00:08:07.747652125 +0000 UTC m=+5.645026775,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.813996 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6deb21e6129 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:08.027218217 +0000 UTC m=+5.924592857,LastTimestamp:2026-03-18 00:08:08.027218217 +0000 UTC m=+5.924592857,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.820144 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6deb3353f66 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:08.045494118 +0000 UTC m=+5.942868758,LastTimestamp:2026-03-18 00:08:08.045494118 +0000 UTC m=+5.942868758,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.825307 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6deb34ebcec openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:08.047164652 +0000 UTC m=+5.944539292,LastTimestamp:2026-03-18 00:08:08.047164652 +0000 UTC m=+5.944539292,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.831811 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6dec3b851b6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:08.322519478 +0000 UTC m=+6.219894128,LastTimestamp:2026-03-18 00:08:08.322519478 +0000 UTC m=+6.219894128,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.838751 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189dc6dec4c42715 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:08.340072213 +0000 UTC m=+6.237446863,LastTimestamp:2026-03-18 00:08:08.340072213 +0000 UTC m=+6.237446863,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.848362 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 18 00:08:57 crc kubenswrapper[4983]: &Event{ObjectMeta:{kube-apiserver-crc.189dc6e0d951ebd0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 18 00:08:57 crc kubenswrapper[4983]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 18 00:08:57 crc kubenswrapper[4983]: Mar 18 00:08:57 crc kubenswrapper[4983]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:17.274842064 +0000 UTC m=+15.172216674,LastTimestamp:2026-03-18 00:08:17.274842064 +0000 UTC m=+15.172216674,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 00:08:57 crc kubenswrapper[4983]: > Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.853999 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6e0d952f4be openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:17.274909886 +0000 UTC m=+15.172284506,LastTimestamp:2026-03-18 00:08:17.274909886 +0000 UTC m=+15.172284506,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.861263 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189dc6e0d951ebd0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 18 00:08:57 crc kubenswrapper[4983]: &Event{ObjectMeta:{kube-apiserver-crc.189dc6e0d951ebd0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 18 00:08:57 crc kubenswrapper[4983]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 18 00:08:57 crc kubenswrapper[4983]: Mar 18 00:08:57 crc kubenswrapper[4983]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:17.274842064 +0000 UTC m=+15.172216674,LastTimestamp:2026-03-18 00:08:17.280733171 +0000 UTC m=+15.178107811,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 00:08:57 crc kubenswrapper[4983]: > Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.867757 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189dc6e0d952f4be\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6e0d952f4be openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:17.274909886 +0000 UTC m=+15.172284506,LastTimestamp:2026-03-18 00:08:17.280948587 +0000 UTC m=+15.178323217,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.873986 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 18 00:08:57 crc kubenswrapper[4983]: &Event{ObjectMeta:{kube-apiserver-crc.189dc6e0dea8ed9f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Liveness probe error: Get "https://192.168.126.11:17697/healthz": dial tcp 192.168.126.11:17697: connect: connection refused Mar 18 00:08:57 crc kubenswrapper[4983]: body: Mar 18 00:08:57 crc kubenswrapper[4983]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:17.364430239 +0000 UTC m=+15.261804849,LastTimestamp:2026-03-18 00:08:17.364430239 +0000 UTC m=+15.261804849,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 00:08:57 crc kubenswrapper[4983]: > Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.880788 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6e0dea9becf openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Liveness probe failed: Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:17.364483791 +0000 UTC m=+15.261858401,LastTimestamp:2026-03-18 00:08:17.364483791 +0000 UTC m=+15.261858401,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.888168 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189dc6de292c3b16\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189dc6de292c3b16 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:05.729647382 +0000 UTC m=+3.627021992,LastTimestamp:2026-03-18 00:08:17.978067032 +0000 UTC m=+15.875441692,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.894982 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 00:08:57 crc kubenswrapper[4983]: &Event{ObjectMeta:{kube-controller-manager-crc.189dc6e11973734a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 18 00:08:57 crc kubenswrapper[4983]: body: Mar 18 00:08:57 crc kubenswrapper[4983]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:18.350781258 +0000 UTC m=+16.248155868,LastTimestamp:2026-03-18 00:08:18.350781258 +0000 UTC m=+16.248155868,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 00:08:57 crc kubenswrapper[4983]: > Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.901374 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6e119742345 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:18.350826309 +0000 UTC m=+16.248200919,LastTimestamp:2026-03-18 00:08:18.350826309 +0000 UTC m=+16.248200919,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.910921 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189dc6e11973734a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 00:08:57 crc kubenswrapper[4983]: &Event{ObjectMeta:{kube-controller-manager-crc.189dc6e11973734a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 18 00:08:57 crc kubenswrapper[4983]: body: Mar 18 00:08:57 crc kubenswrapper[4983]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:18.350781258 +0000 UTC m=+16.248155868,LastTimestamp:2026-03-18 00:08:28.350905169 +0000 UTC m=+26.248279809,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 00:08:57 crc kubenswrapper[4983]: > Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.918020 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189dc6e119742345\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6e119742345 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:18.350826309 +0000 UTC m=+16.248200919,LastTimestamp:2026-03-18 00:08:28.350988302 +0000 UTC m=+26.248362952,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.924803 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 00:08:57 crc kubenswrapper[4983]: &Event{ObjectMeta:{kube-controller-manager-crc.189dc6e5126490ae openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": read tcp 192.168.126.11:48610->192.168.126.11:10357: read: connection reset by peer Mar 18 00:08:57 crc kubenswrapper[4983]: body: Mar 18 00:08:57 crc kubenswrapper[4983]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:35.412234414 +0000 UTC m=+33.309609064,LastTimestamp:2026-03-18 00:08:35.412234414 +0000 UTC m=+33.309609064,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 00:08:57 crc kubenswrapper[4983]: > Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.932041 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6e5126578a4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:48610->192.168.126.11:10357: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:35.412293796 +0000 UTC m=+33.309668446,LastTimestamp:2026-03-18 00:08:35.412293796 +0000 UTC m=+33.309668446,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.936175 4983 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6e5129724a0 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:35.415549088 +0000 UTC m=+33.312923738,LastTimestamp:2026-03-18 00:08:35.415549088 +0000 UTC m=+33.312923738,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.940001 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189dc6ddc12b7f42\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6ddc12b7f42 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:03.984768834 +0000 UTC m=+1.882143484,LastTimestamp:2026-03-18 00:08:35.438128977 +0000 UTC m=+33.335503577,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.945813 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189dc6ddd5493528\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6ddd5493528 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.322260264 +0000 UTC m=+2.219634904,LastTimestamp:2026-03-18 00:08:35.678878139 +0000 UTC m=+33.576252749,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.952603 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189dc6ddd61f6129\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6ddd61f6129 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:04.336296233 +0000 UTC m=+2.233670883,LastTimestamp:2026-03-18 00:08:35.692579917 +0000 UTC m=+33.589954527,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.961991 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189dc6e11973734a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 00:08:57 crc kubenswrapper[4983]: &Event{ObjectMeta:{kube-controller-manager-crc.189dc6e11973734a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 18 00:08:57 crc kubenswrapper[4983]: body: Mar 18 00:08:57 crc kubenswrapper[4983]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:18.350781258 +0000 UTC m=+16.248155868,LastTimestamp:2026-03-18 00:08:48.35106588 +0000 UTC m=+46.248440540,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 00:08:57 crc kubenswrapper[4983]: > Mar 18 00:08:57 crc kubenswrapper[4983]: E0318 00:08:57.969263 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189dc6e119742345\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189dc6e119742345 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:18.350826309 +0000 UTC m=+16.248200919,LastTimestamp:2026-03-18 00:08:48.351302547 +0000 UTC m=+46.248677187,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:08:58 crc kubenswrapper[4983]: I0318 00:08:58.125319 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 00:08:58 crc kubenswrapper[4983]: I0318 00:08:58.125508 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:58 crc kubenswrapper[4983]: I0318 00:08:58.126640 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:58 crc kubenswrapper[4983]: I0318 00:08:58.126691 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:58 crc kubenswrapper[4983]: I0318 00:08:58.126708 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:58 crc kubenswrapper[4983]: I0318 00:08:58.350480 4983 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 00:08:58 crc kubenswrapper[4983]: I0318 00:08:58.350660 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 00:08:58 crc kubenswrapper[4983]: E0318 00:08:58.359051 4983 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189dc6e11973734a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 00:08:58 crc kubenswrapper[4983]: &Event{ObjectMeta:{kube-controller-manager-crc.189dc6e11973734a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 18 00:08:58 crc kubenswrapper[4983]: body: Mar 18 00:08:58 crc kubenswrapper[4983]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:08:18.350781258 +0000 UTC m=+16.248155868,LastTimestamp:2026-03-18 00:08:58.350606138 +0000 UTC m=+56.247980778,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 00:08:58 crc kubenswrapper[4983]: > Mar 18 00:08:58 crc kubenswrapper[4983]: E0318 00:08:58.710483 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 00:08:58 crc kubenswrapper[4983]: I0318 00:08:58.711704 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:08:58 crc kubenswrapper[4983]: I0318 00:08:58.713377 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:08:58 crc kubenswrapper[4983]: I0318 00:08:58.713428 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:08:58 crc kubenswrapper[4983]: I0318 00:08:58.713448 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:08:58 crc kubenswrapper[4983]: I0318 00:08:58.713481 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:08:58 crc kubenswrapper[4983]: E0318 00:08:58.720366 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 00:08:58 crc kubenswrapper[4983]: I0318 00:08:58.760994 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:08:59 crc kubenswrapper[4983]: I0318 00:08:59.759847 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:00 crc kubenswrapper[4983]: I0318 00:09:00.760429 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:01 crc kubenswrapper[4983]: I0318 00:09:01.761977 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:02 crc kubenswrapper[4983]: I0318 00:09:02.760035 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:02 crc kubenswrapper[4983]: I0318 00:09:02.844081 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:09:02 crc kubenswrapper[4983]: I0318 00:09:02.845605 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:02 crc kubenswrapper[4983]: I0318 00:09:02.845661 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:02 crc kubenswrapper[4983]: I0318 00:09:02.845678 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:02 crc kubenswrapper[4983]: I0318 00:09:02.846620 4983 scope.go:117] "RemoveContainer" containerID="3ae69de2b7a7acff7586d0220e4f2487998f8552d7a6b1fb82efe01a56c9b279" Mar 18 00:09:02 crc kubenswrapper[4983]: E0318 00:09:02.929241 4983 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 00:09:03 crc kubenswrapper[4983]: I0318 00:09:03.146107 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 18 00:09:03 crc kubenswrapper[4983]: I0318 00:09:03.148470 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626"} Mar 18 00:09:03 crc kubenswrapper[4983]: I0318 00:09:03.148620 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:09:03 crc kubenswrapper[4983]: I0318 00:09:03.149702 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:03 crc kubenswrapper[4983]: I0318 00:09:03.149979 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:03 crc kubenswrapper[4983]: I0318 00:09:03.150004 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:03 crc kubenswrapper[4983]: I0318 00:09:03.759571 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:04 crc kubenswrapper[4983]: I0318 00:09:04.152752 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 18 00:09:04 crc kubenswrapper[4983]: I0318 00:09:04.154800 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 18 00:09:04 crc kubenswrapper[4983]: I0318 00:09:04.157230 4983 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626" exitCode=255 Mar 18 00:09:04 crc kubenswrapper[4983]: I0318 00:09:04.157288 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626"} Mar 18 00:09:04 crc kubenswrapper[4983]: I0318 00:09:04.157343 4983 scope.go:117] "RemoveContainer" containerID="3ae69de2b7a7acff7586d0220e4f2487998f8552d7a6b1fb82efe01a56c9b279" Mar 18 00:09:04 crc kubenswrapper[4983]: I0318 00:09:04.157676 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:09:04 crc kubenswrapper[4983]: I0318 00:09:04.164229 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:04 crc kubenswrapper[4983]: I0318 00:09:04.164303 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:04 crc kubenswrapper[4983]: I0318 00:09:04.164321 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:04 crc kubenswrapper[4983]: I0318 00:09:04.165303 4983 scope.go:117] "RemoveContainer" containerID="b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626" Mar 18 00:09:04 crc kubenswrapper[4983]: E0318 00:09:04.165604 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:09:04 crc kubenswrapper[4983]: I0318 00:09:04.758024 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:05 crc kubenswrapper[4983]: I0318 00:09:05.161849 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 18 00:09:05 crc kubenswrapper[4983]: I0318 00:09:05.356028 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:09:05 crc kubenswrapper[4983]: I0318 00:09:05.356183 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:09:05 crc kubenswrapper[4983]: I0318 00:09:05.360063 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:05 crc kubenswrapper[4983]: I0318 00:09:05.360106 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:05 crc kubenswrapper[4983]: I0318 00:09:05.360117 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:05 crc kubenswrapper[4983]: I0318 00:09:05.366383 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:09:05 crc kubenswrapper[4983]: I0318 00:09:05.720889 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:09:05 crc kubenswrapper[4983]: I0318 00:09:05.722587 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:05 crc kubenswrapper[4983]: I0318 00:09:05.722631 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:05 crc kubenswrapper[4983]: I0318 00:09:05.722649 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:05 crc kubenswrapper[4983]: I0318 00:09:05.722681 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:09:05 crc kubenswrapper[4983]: E0318 00:09:05.725787 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 00:09:05 crc kubenswrapper[4983]: E0318 00:09:05.727033 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 00:09:05 crc kubenswrapper[4983]: I0318 00:09:05.760026 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:06 crc kubenswrapper[4983]: I0318 00:09:06.166852 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:09:06 crc kubenswrapper[4983]: I0318 00:09:06.168201 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:06 crc kubenswrapper[4983]: I0318 00:09:06.168249 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:06 crc kubenswrapper[4983]: I0318 00:09:06.168263 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:06 crc kubenswrapper[4983]: I0318 00:09:06.760360 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:07 crc kubenswrapper[4983]: I0318 00:09:07.363466 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:09:07 crc kubenswrapper[4983]: I0318 00:09:07.363606 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:09:07 crc kubenswrapper[4983]: I0318 00:09:07.364690 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:07 crc kubenswrapper[4983]: I0318 00:09:07.364750 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:07 crc kubenswrapper[4983]: I0318 00:09:07.364768 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:07 crc kubenswrapper[4983]: I0318 00:09:07.365538 4983 scope.go:117] "RemoveContainer" containerID="b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626" Mar 18 00:09:07 crc kubenswrapper[4983]: E0318 00:09:07.366066 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:09:07 crc kubenswrapper[4983]: I0318 00:09:07.758754 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:08 crc kubenswrapper[4983]: I0318 00:09:08.758971 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:09 crc kubenswrapper[4983]: I0318 00:09:09.757493 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:10 crc kubenswrapper[4983]: I0318 00:09:10.760425 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:10 crc kubenswrapper[4983]: I0318 00:09:10.821202 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:09:10 crc kubenswrapper[4983]: I0318 00:09:10.821357 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:09:10 crc kubenswrapper[4983]: I0318 00:09:10.823320 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:10 crc kubenswrapper[4983]: I0318 00:09:10.823347 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:10 crc kubenswrapper[4983]: I0318 00:09:10.823354 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:10 crc kubenswrapper[4983]: I0318 00:09:10.823805 4983 scope.go:117] "RemoveContainer" containerID="b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626" Mar 18 00:09:10 crc kubenswrapper[4983]: E0318 00:09:10.823960 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:09:11 crc kubenswrapper[4983]: I0318 00:09:11.759298 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:12 crc kubenswrapper[4983]: I0318 00:09:12.727294 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:09:12 crc kubenswrapper[4983]: I0318 00:09:12.728660 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:12 crc kubenswrapper[4983]: I0318 00:09:12.728713 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:12 crc kubenswrapper[4983]: I0318 00:09:12.728731 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:12 crc kubenswrapper[4983]: I0318 00:09:12.728793 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:09:12 crc kubenswrapper[4983]: E0318 00:09:12.732710 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 00:09:12 crc kubenswrapper[4983]: E0318 00:09:12.733089 4983 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 00:09:12 crc kubenswrapper[4983]: I0318 00:09:12.754400 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:12 crc kubenswrapper[4983]: E0318 00:09:12.929397 4983 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 00:09:13 crc kubenswrapper[4983]: I0318 00:09:13.760757 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:14 crc kubenswrapper[4983]: I0318 00:09:14.712471 4983 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 00:09:14 crc kubenswrapper[4983]: I0318 00:09:14.730169 4983 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 18 00:09:14 crc kubenswrapper[4983]: I0318 00:09:14.759301 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:15 crc kubenswrapper[4983]: I0318 00:09:15.761308 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:16 crc kubenswrapper[4983]: I0318 00:09:16.758992 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:17 crc kubenswrapper[4983]: I0318 00:09:17.761435 4983 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 00:09:17 crc kubenswrapper[4983]: I0318 00:09:17.959084 4983 csr.go:261] certificate signing request csr-xxm8j is approved, waiting to be issued Mar 18 00:09:17 crc kubenswrapper[4983]: I0318 00:09:17.970779 4983 csr.go:257] certificate signing request csr-xxm8j is issued Mar 18 00:09:18 crc kubenswrapper[4983]: I0318 00:09:18.009615 4983 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 18 00:09:18 crc kubenswrapper[4983]: I0318 00:09:18.582200 4983 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 18 00:09:18 crc kubenswrapper[4983]: I0318 00:09:18.972638 4983 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-01 04:15:09.84623967 +0000 UTC Mar 18 00:09:18 crc kubenswrapper[4983]: I0318 00:09:18.972687 4983 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6940h5m50.873556102s for next certificate rotation Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.733979 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.735804 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.735850 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.735868 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.736064 4983 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.746207 4983 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.746518 4983 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 18 00:09:19 crc kubenswrapper[4983]: E0318 00:09:19.746553 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.751667 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.751870 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.752035 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.752171 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.752306 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:19Z","lastTransitionTime":"2026-03-18T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:19 crc kubenswrapper[4983]: E0318 00:09:19.773248 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.784054 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.784113 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.784130 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.784156 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.784173 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:19Z","lastTransitionTime":"2026-03-18T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:19 crc kubenswrapper[4983]: E0318 00:09:19.800103 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.811569 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.811848 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.812079 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.812231 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.812360 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:19Z","lastTransitionTime":"2026-03-18T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:19 crc kubenswrapper[4983]: E0318 00:09:19.832453 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.843075 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.843127 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.843146 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.843170 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:19 crc kubenswrapper[4983]: I0318 00:09:19.843187 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:19Z","lastTransitionTime":"2026-03-18T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:19 crc kubenswrapper[4983]: E0318 00:09:19.858757 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:19 crc kubenswrapper[4983]: E0318 00:09:19.859021 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 00:09:19 crc kubenswrapper[4983]: E0318 00:09:19.859072 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:19 crc kubenswrapper[4983]: E0318 00:09:19.959598 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:20 crc kubenswrapper[4983]: E0318 00:09:20.060398 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:20 crc kubenswrapper[4983]: E0318 00:09:20.161595 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:20 crc kubenswrapper[4983]: E0318 00:09:20.262133 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:20 crc kubenswrapper[4983]: E0318 00:09:20.362648 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:20 crc kubenswrapper[4983]: E0318 00:09:20.463372 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:20 crc kubenswrapper[4983]: E0318 00:09:20.564110 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:20 crc kubenswrapper[4983]: E0318 00:09:20.664431 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:20 crc kubenswrapper[4983]: E0318 00:09:20.765443 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:20 crc kubenswrapper[4983]: I0318 00:09:20.844159 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:09:20 crc kubenswrapper[4983]: I0318 00:09:20.845509 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:20 crc kubenswrapper[4983]: I0318 00:09:20.845567 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:20 crc kubenswrapper[4983]: I0318 00:09:20.845584 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:20 crc kubenswrapper[4983]: E0318 00:09:20.866292 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:20 crc kubenswrapper[4983]: E0318 00:09:20.966430 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:21 crc kubenswrapper[4983]: E0318 00:09:21.067330 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:21 crc kubenswrapper[4983]: E0318 00:09:21.168111 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:21 crc kubenswrapper[4983]: E0318 00:09:21.268300 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:21 crc kubenswrapper[4983]: E0318 00:09:21.369455 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:21 crc kubenswrapper[4983]: E0318 00:09:21.470269 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:21 crc kubenswrapper[4983]: E0318 00:09:21.570642 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:21 crc kubenswrapper[4983]: E0318 00:09:21.671077 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:21 crc kubenswrapper[4983]: E0318 00:09:21.771682 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:21 crc kubenswrapper[4983]: E0318 00:09:21.872617 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:21 crc kubenswrapper[4983]: E0318 00:09:21.972961 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:22 crc kubenswrapper[4983]: E0318 00:09:22.073821 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:22 crc kubenswrapper[4983]: E0318 00:09:22.174429 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:22 crc kubenswrapper[4983]: E0318 00:09:22.275072 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:22 crc kubenswrapper[4983]: E0318 00:09:22.375625 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:22 crc kubenswrapper[4983]: E0318 00:09:22.476644 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:22 crc kubenswrapper[4983]: E0318 00:09:22.577004 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:22 crc kubenswrapper[4983]: E0318 00:09:22.677616 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:22 crc kubenswrapper[4983]: I0318 00:09:22.695872 4983 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 18 00:09:22 crc kubenswrapper[4983]: E0318 00:09:22.778650 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:22 crc kubenswrapper[4983]: E0318 00:09:22.878809 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:22 crc kubenswrapper[4983]: E0318 00:09:22.929707 4983 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 00:09:22 crc kubenswrapper[4983]: E0318 00:09:22.980060 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:23 crc kubenswrapper[4983]: E0318 00:09:23.081006 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:23 crc kubenswrapper[4983]: E0318 00:09:23.182466 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:23 crc kubenswrapper[4983]: E0318 00:09:23.283567 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:23 crc kubenswrapper[4983]: E0318 00:09:23.384692 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:23 crc kubenswrapper[4983]: E0318 00:09:23.485726 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:23 crc kubenswrapper[4983]: E0318 00:09:23.591837 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:23 crc kubenswrapper[4983]: E0318 00:09:23.692389 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:23 crc kubenswrapper[4983]: E0318 00:09:23.792675 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:23 crc kubenswrapper[4983]: E0318 00:09:23.892831 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:23 crc kubenswrapper[4983]: E0318 00:09:23.993279 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:24 crc kubenswrapper[4983]: E0318 00:09:24.094475 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:24 crc kubenswrapper[4983]: E0318 00:09:24.194897 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:24 crc kubenswrapper[4983]: E0318 00:09:24.296388 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:24 crc kubenswrapper[4983]: E0318 00:09:24.396763 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:24 crc kubenswrapper[4983]: E0318 00:09:24.497818 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:24 crc kubenswrapper[4983]: E0318 00:09:24.599170 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:24 crc kubenswrapper[4983]: E0318 00:09:24.699905 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:24 crc kubenswrapper[4983]: E0318 00:09:24.800911 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:24 crc kubenswrapper[4983]: E0318 00:09:24.901878 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:25 crc kubenswrapper[4983]: E0318 00:09:25.002338 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:25 crc kubenswrapper[4983]: E0318 00:09:25.102484 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:25 crc kubenswrapper[4983]: E0318 00:09:25.203979 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:25 crc kubenswrapper[4983]: E0318 00:09:25.304308 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:25 crc kubenswrapper[4983]: E0318 00:09:25.404853 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:25 crc kubenswrapper[4983]: E0318 00:09:25.505844 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:25 crc kubenswrapper[4983]: E0318 00:09:25.606077 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:25 crc kubenswrapper[4983]: E0318 00:09:25.706473 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:25 crc kubenswrapper[4983]: E0318 00:09:25.807191 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:25 crc kubenswrapper[4983]: I0318 00:09:25.844286 4983 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 00:09:25 crc kubenswrapper[4983]: I0318 00:09:25.845964 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:25 crc kubenswrapper[4983]: I0318 00:09:25.846021 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:25 crc kubenswrapper[4983]: I0318 00:09:25.846038 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:25 crc kubenswrapper[4983]: I0318 00:09:25.847069 4983 scope.go:117] "RemoveContainer" containerID="b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626" Mar 18 00:09:25 crc kubenswrapper[4983]: E0318 00:09:25.847354 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:09:25 crc kubenswrapper[4983]: E0318 00:09:25.907889 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:26 crc kubenswrapper[4983]: E0318 00:09:26.008039 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:26 crc kubenswrapper[4983]: E0318 00:09:26.109048 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:26 crc kubenswrapper[4983]: E0318 00:09:26.210103 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:26 crc kubenswrapper[4983]: E0318 00:09:26.310204 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:26 crc kubenswrapper[4983]: E0318 00:09:26.410517 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:26 crc kubenswrapper[4983]: E0318 00:09:26.511425 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:26 crc kubenswrapper[4983]: E0318 00:09:26.611551 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:26 crc kubenswrapper[4983]: E0318 00:09:26.711738 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:26 crc kubenswrapper[4983]: E0318 00:09:26.811996 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:26 crc kubenswrapper[4983]: E0318 00:09:26.913222 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.013411 4983 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.113612 4983 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.116437 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.116496 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.116518 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.116546 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.116567 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:27Z","lastTransitionTime":"2026-03-18T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.220115 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.220249 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.220276 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.220306 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.220325 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:27Z","lastTransitionTime":"2026-03-18T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.323364 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.323421 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.323441 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.323463 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.323481 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:27Z","lastTransitionTime":"2026-03-18T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.426452 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.426510 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.426527 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.426552 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.426570 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:27Z","lastTransitionTime":"2026-03-18T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.529481 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.529524 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.529539 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.529561 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.529577 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:27Z","lastTransitionTime":"2026-03-18T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.632685 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.633065 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.633276 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.633425 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.633566 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:27Z","lastTransitionTime":"2026-03-18T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.736806 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.736861 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.736879 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.736900 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.736916 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:27Z","lastTransitionTime":"2026-03-18T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.780719 4983 apiserver.go:52] "Watching apiserver" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.786640 4983 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.787048 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.787530 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.787715 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.787729 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.787905 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.788209 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.788342 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.788581 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.788686 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.789073 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.792761 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.793053 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.792806 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.792795 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.792987 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.793691 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.794259 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.794556 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.795464 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.838742 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.841349 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.841410 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.841487 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.841517 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.841535 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:27Z","lastTransitionTime":"2026-03-18T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.857105 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.864654 4983 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.877030 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.893062 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.908435 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.916847 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.916964 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917025 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917078 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917126 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917175 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917228 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917281 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917326 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917375 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917422 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917470 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917520 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917564 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917613 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917657 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917708 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917762 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917796 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917807 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.917814 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918000 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918057 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918092 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918228 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918307 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918420 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918454 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918488 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918520 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918554 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918586 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918618 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918651 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918681 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918715 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918747 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918778 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918821 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918869 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918916 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918994 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919037 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919078 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919126 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919172 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919204 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919236 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919270 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919301 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919331 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919364 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919397 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919432 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919464 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919493 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919530 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919567 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919599 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919631 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919663 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919695 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919727 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919758 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919792 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919827 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919861 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919893 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919958 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919992 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920026 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920059 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920091 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920125 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920182 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920215 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920248 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920280 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920310 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918115 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918321 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920359 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918618 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920390 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920429 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920464 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920497 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920545 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920577 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920610 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920643 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920675 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920709 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920740 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920773 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920806 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920841 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920889 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.921016 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.922652 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.922781 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.922888 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.922951 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.922984 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923019 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923051 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923085 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923118 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923157 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923196 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923237 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923284 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923324 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923367 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923407 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923446 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923489 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923525 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923559 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923595 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923628 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923661 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923694 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923729 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923762 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923795 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923831 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923866 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923898 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923955 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923992 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924024 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924058 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924091 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924124 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924157 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924191 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924222 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924257 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924291 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924324 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924357 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924392 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924426 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924499 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924537 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924573 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924605 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924646 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924723 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924791 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924884 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.924977 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925016 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925050 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925087 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925122 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925155 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925189 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925223 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925257 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925290 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925328 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925363 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925403 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925437 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925470 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925504 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925544 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925585 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925617 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925650 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925685 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925721 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925756 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925792 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925829 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925986 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918768 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926025 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926061 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926103 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926138 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926192 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926228 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926268 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926301 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926336 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926371 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926406 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926441 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926408 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926476 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926513 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926552 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926588 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926629 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926666 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926706 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926742 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926778 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926811 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926846 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926883 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926979 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927024 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927066 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927106 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927153 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927189 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927232 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927271 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927307 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927346 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927388 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927424 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927469 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927509 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927590 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927615 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927637 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927659 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927681 4983 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927700 4983 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.939526 4983 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.941472 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.944203 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.945993 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.947768 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.950098 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.950138 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.950154 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.950220 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.950240 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:27Z","lastTransitionTime":"2026-03-18T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.950249 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.951915 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.918785 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919223 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919277 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919365 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919403 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919411 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919863 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919918 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919906 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.919921 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920179 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920312 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920255 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920651 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920600 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920779 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.920836 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.921017 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.921137 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.921157 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.921296 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.921332 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.921525 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.921625 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.921836 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.922285 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.922617 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923045 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923158 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923323 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923336 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923634 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923767 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.923797 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925771 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.925997 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926068 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.926627 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927250 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927280 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927436 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927769 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.927781 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.929452 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.929460 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.929740 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.929962 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.930152 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.930230 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.930495 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.930803 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.931976 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.932139 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.932235 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.934607 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.935005 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.935393 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.935647 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.935944 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.936227 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.936576 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.937241 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.937419 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.937648 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.937734 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.937740 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.937877 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.938243 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.938345 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.938402 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.938624 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.938799 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.938844 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.938914 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.939335 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.939359 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.939418 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.939455 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.939573 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.939862 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.939901 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.939957 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.940050 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.940095 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.940252 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.940605 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.940690 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.940822 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.940913 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.941316 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.941737 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.941770 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.941784 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.941846 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.941848 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.942370 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.943433 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.943474 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.943582 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.944081 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.944157 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.944364 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.945110 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.945497 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.945526 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.945653 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.946061 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.946143 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.947181 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.947424 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.947719 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.947999 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.948018 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.948239 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.948907 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.949304 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.949663 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.950423 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.950671 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.951088 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.951259 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:09:28.451227604 +0000 UTC m=+86.348602244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.951421 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.951572 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.951565 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.951729 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.952195 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.952403 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.952560 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.952616 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.952872 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.953180 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.953413 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.954021 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.954524 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.954660 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:28.454638192 +0000 UTC m=+86.352012842 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.954969 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.956273 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.965513 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.967819 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.968112 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.958357 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.968645 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:28.468611903 +0000 UTC m=+86.365986623 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.968679 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.968961 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.968990 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.969071 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:28.469055125 +0000 UTC m=+86.366429845 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.969108 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.969186 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.969501 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.969354 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.969993 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.970019 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.970039 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:27 crc kubenswrapper[4983]: E0318 00:09:27.970104 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:28.470081375 +0000 UTC m=+86.367456015 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.970257 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.970505 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.970582 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.970300 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.970699 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.970831 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.970983 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.971035 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.973357 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.973705 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.973806 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.975485 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.975526 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.975556 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.975729 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.975809 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.975962 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.975988 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.976501 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.977104 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.977209 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.977260 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.977313 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.977347 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.977401 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.977488 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.977632 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.977867 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.978226 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.981043 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.981873 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.982385 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.984193 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.986424 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.986554 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.986855 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.986982 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.987291 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.987301 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.987887 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.988014 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.988760 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.989075 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.989493 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:09:27 crc kubenswrapper[4983]: I0318 00:09:27.998515 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.008757 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.011398 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.012128 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029029 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029385 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029183 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029508 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029533 4983 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029553 4983 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029573 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029592 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029612 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029630 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029648 4983 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029670 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029676 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029689 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029792 4983 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029813 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029833 4983 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029853 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029870 4983 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029888 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029906 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029955 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029976 4983 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.029996 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030015 4983 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030032 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030050 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030067 4983 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030086 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030106 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030124 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030142 4983 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030160 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030211 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030242 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030267 4983 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030290 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030314 4983 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030338 4983 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030362 4983 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030387 4983 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030412 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030436 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030462 4983 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030486 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030511 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030535 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030560 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030585 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030605 4983 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030626 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030644 4983 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030664 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030683 4983 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030703 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030723 4983 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030742 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030761 4983 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030779 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030798 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030818 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030837 4983 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030855 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030876 4983 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030894 4983 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030915 4983 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030965 4983 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.030989 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031015 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031039 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031063 4983 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031088 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031113 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031139 4983 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031166 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031192 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031217 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031243 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031269 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031293 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031378 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031402 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031428 4983 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031457 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031482 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031506 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031535 4983 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031562 4983 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031588 4983 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031610 4983 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031628 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031648 4983 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031668 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031686 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031702 4983 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031718 4983 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031736 4983 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031752 4983 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031910 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031960 4983 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031980 4983 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.031997 4983 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032014 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032034 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032050 4983 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032066 4983 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032084 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032102 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032120 4983 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032139 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032158 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032180 4983 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032198 4983 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032220 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032246 4983 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032270 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032319 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032336 4983 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032354 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032377 4983 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032399 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032423 4983 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032448 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032471 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032490 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032508 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032527 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032545 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032562 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032579 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032596 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032613 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032630 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032648 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032670 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032693 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032713 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032733 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032757 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032777 4983 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032795 4983 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032827 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032863 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032885 4983 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032903 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032920 4983 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032971 4983 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.032989 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033006 4983 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033022 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033039 4983 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033056 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033074 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033092 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033111 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033165 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033183 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033199 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033219 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033236 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033253 4983 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033270 4983 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033286 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033303 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033320 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033336 4983 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033353 4983 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033369 4983 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033385 4983 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033402 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033419 4983 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033435 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033452 4983 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033468 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033485 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033503 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033520 4983 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033538 4983 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033556 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033574 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033615 4983 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033633 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033650 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033667 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033686 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033704 4983 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033720 4983 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033738 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033755 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033774 4983 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033793 4983 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.033820 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.056672 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.056703 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.056720 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.056740 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.056752 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:28Z","lastTransitionTime":"2026-03-18T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.121633 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.131423 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 00:09:28 crc kubenswrapper[4983]: W0318 00:09:28.144169 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-6cf5fc83bcf8b1197b97bb6cd676f211ee4a6104d204e0a30f6eca18a563cc1b WatchSource:0}: Error finding container 6cf5fc83bcf8b1197b97bb6cd676f211ee4a6104d204e0a30f6eca18a563cc1b: Status 404 returned error can't find the container with id 6cf5fc83bcf8b1197b97bb6cd676f211ee4a6104d204e0a30f6eca18a563cc1b Mar 18 00:09:28 crc kubenswrapper[4983]: W0318 00:09:28.147606 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-4a196b0ea535d09521d594621cbd2c782b2a324baff5a2a12fd4de5c0ac50b98 WatchSource:0}: Error finding container 4a196b0ea535d09521d594621cbd2c782b2a324baff5a2a12fd4de5c0ac50b98: Status 404 returned error can't find the container with id 4a196b0ea535d09521d594621cbd2c782b2a324baff5a2a12fd4de5c0ac50b98 Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.148871 4983 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 00:09:28 crc kubenswrapper[4983]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 18 00:09:28 crc kubenswrapper[4983]: set -o allexport Mar 18 00:09:28 crc kubenswrapper[4983]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 18 00:09:28 crc kubenswrapper[4983]: source /etc/kubernetes/apiserver-url.env Mar 18 00:09:28 crc kubenswrapper[4983]: else Mar 18 00:09:28 crc kubenswrapper[4983]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 18 00:09:28 crc kubenswrapper[4983]: exit 1 Mar 18 00:09:28 crc kubenswrapper[4983]: fi Mar 18 00:09:28 crc kubenswrapper[4983]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 18 00:09:28 crc kubenswrapper[4983]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 00:09:28 crc kubenswrapper[4983]: > logger="UnhandledError" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.149258 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.149360 4983 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 00:09:28 crc kubenswrapper[4983]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 18 00:09:28 crc kubenswrapper[4983]: if [[ -f "/env/_master" ]]; then Mar 18 00:09:28 crc kubenswrapper[4983]: set -o allexport Mar 18 00:09:28 crc kubenswrapper[4983]: source "/env/_master" Mar 18 00:09:28 crc kubenswrapper[4983]: set +o allexport Mar 18 00:09:28 crc kubenswrapper[4983]: fi Mar 18 00:09:28 crc kubenswrapper[4983]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 18 00:09:28 crc kubenswrapper[4983]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 18 00:09:28 crc kubenswrapper[4983]: ho_enable="--enable-hybrid-overlay" Mar 18 00:09:28 crc kubenswrapper[4983]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 18 00:09:28 crc kubenswrapper[4983]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 18 00:09:28 crc kubenswrapper[4983]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 18 00:09:28 crc kubenswrapper[4983]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 18 00:09:28 crc kubenswrapper[4983]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 18 00:09:28 crc kubenswrapper[4983]: --webhook-host=127.0.0.1 \ Mar 18 00:09:28 crc kubenswrapper[4983]: --webhook-port=9743 \ Mar 18 00:09:28 crc kubenswrapper[4983]: ${ho_enable} \ Mar 18 00:09:28 crc kubenswrapper[4983]: --enable-interconnect \ Mar 18 00:09:28 crc kubenswrapper[4983]: --disable-approver \ Mar 18 00:09:28 crc kubenswrapper[4983]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 18 00:09:28 crc kubenswrapper[4983]: --wait-for-kubernetes-api=200s \ Mar 18 00:09:28 crc kubenswrapper[4983]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 18 00:09:28 crc kubenswrapper[4983]: --loglevel="${LOGLEVEL}" Mar 18 00:09:28 crc kubenswrapper[4983]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 00:09:28 crc kubenswrapper[4983]: > logger="UnhandledError" Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.150703 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.153429 4983 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 00:09:28 crc kubenswrapper[4983]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 18 00:09:28 crc kubenswrapper[4983]: if [[ -f "/env/_master" ]]; then Mar 18 00:09:28 crc kubenswrapper[4983]: set -o allexport Mar 18 00:09:28 crc kubenswrapper[4983]: source "/env/_master" Mar 18 00:09:28 crc kubenswrapper[4983]: set +o allexport Mar 18 00:09:28 crc kubenswrapper[4983]: fi Mar 18 00:09:28 crc kubenswrapper[4983]: Mar 18 00:09:28 crc kubenswrapper[4983]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 18 00:09:28 crc kubenswrapper[4983]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 18 00:09:28 crc kubenswrapper[4983]: --disable-webhook \ Mar 18 00:09:28 crc kubenswrapper[4983]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 18 00:09:28 crc kubenswrapper[4983]: --loglevel="${LOGLEVEL}" Mar 18 00:09:28 crc kubenswrapper[4983]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 00:09:28 crc kubenswrapper[4983]: > logger="UnhandledError" Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.154702 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.157912 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.157958 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.157970 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.157990 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.158005 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:28Z","lastTransitionTime":"2026-03-18T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:28 crc kubenswrapper[4983]: W0318 00:09:28.162880 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-0a401a2b3bfd44f507b67139205709f8c93b778cd5608183735a21361adca375 WatchSource:0}: Error finding container 0a401a2b3bfd44f507b67139205709f8c93b778cd5608183735a21361adca375: Status 404 returned error can't find the container with id 0a401a2b3bfd44f507b67139205709f8c93b778cd5608183735a21361adca375 Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.165746 4983 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.167016 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.232180 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6cf5fc83bcf8b1197b97bb6cd676f211ee4a6104d204e0a30f6eca18a563cc1b"} Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.233343 4983 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 00:09:28 crc kubenswrapper[4983]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 18 00:09:28 crc kubenswrapper[4983]: set -o allexport Mar 18 00:09:28 crc kubenswrapper[4983]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 18 00:09:28 crc kubenswrapper[4983]: source /etc/kubernetes/apiserver-url.env Mar 18 00:09:28 crc kubenswrapper[4983]: else Mar 18 00:09:28 crc kubenswrapper[4983]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 18 00:09:28 crc kubenswrapper[4983]: exit 1 Mar 18 00:09:28 crc kubenswrapper[4983]: fi Mar 18 00:09:28 crc kubenswrapper[4983]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 18 00:09:28 crc kubenswrapper[4983]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 00:09:28 crc kubenswrapper[4983]: > logger="UnhandledError" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.233828 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"0a401a2b3bfd44f507b67139205709f8c93b778cd5608183735a21361adca375"} Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.234428 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.236216 4983 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.236462 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4a196b0ea535d09521d594621cbd2c782b2a324baff5a2a12fd4de5c0ac50b98"} Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.237667 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.238422 4983 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 00:09:28 crc kubenswrapper[4983]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 18 00:09:28 crc kubenswrapper[4983]: if [[ -f "/env/_master" ]]; then Mar 18 00:09:28 crc kubenswrapper[4983]: set -o allexport Mar 18 00:09:28 crc kubenswrapper[4983]: source "/env/_master" Mar 18 00:09:28 crc kubenswrapper[4983]: set +o allexport Mar 18 00:09:28 crc kubenswrapper[4983]: fi Mar 18 00:09:28 crc kubenswrapper[4983]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 18 00:09:28 crc kubenswrapper[4983]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 18 00:09:28 crc kubenswrapper[4983]: ho_enable="--enable-hybrid-overlay" Mar 18 00:09:28 crc kubenswrapper[4983]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 18 00:09:28 crc kubenswrapper[4983]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 18 00:09:28 crc kubenswrapper[4983]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 18 00:09:28 crc kubenswrapper[4983]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 18 00:09:28 crc kubenswrapper[4983]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 18 00:09:28 crc kubenswrapper[4983]: --webhook-host=127.0.0.1 \ Mar 18 00:09:28 crc kubenswrapper[4983]: --webhook-port=9743 \ Mar 18 00:09:28 crc kubenswrapper[4983]: ${ho_enable} \ Mar 18 00:09:28 crc kubenswrapper[4983]: --enable-interconnect \ Mar 18 00:09:28 crc kubenswrapper[4983]: --disable-approver \ Mar 18 00:09:28 crc kubenswrapper[4983]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 18 00:09:28 crc kubenswrapper[4983]: --wait-for-kubernetes-api=200s \ Mar 18 00:09:28 crc kubenswrapper[4983]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 18 00:09:28 crc kubenswrapper[4983]: --loglevel="${LOGLEVEL}" Mar 18 00:09:28 crc kubenswrapper[4983]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 00:09:28 crc kubenswrapper[4983]: > logger="UnhandledError" Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.241267 4983 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 00:09:28 crc kubenswrapper[4983]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 18 00:09:28 crc kubenswrapper[4983]: if [[ -f "/env/_master" ]]; then Mar 18 00:09:28 crc kubenswrapper[4983]: set -o allexport Mar 18 00:09:28 crc kubenswrapper[4983]: source "/env/_master" Mar 18 00:09:28 crc kubenswrapper[4983]: set +o allexport Mar 18 00:09:28 crc kubenswrapper[4983]: fi Mar 18 00:09:28 crc kubenswrapper[4983]: Mar 18 00:09:28 crc kubenswrapper[4983]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 18 00:09:28 crc kubenswrapper[4983]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 18 00:09:28 crc kubenswrapper[4983]: --disable-webhook \ Mar 18 00:09:28 crc kubenswrapper[4983]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 18 00:09:28 crc kubenswrapper[4983]: --loglevel="${LOGLEVEL}" Mar 18 00:09:28 crc kubenswrapper[4983]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 18 00:09:28 crc kubenswrapper[4983]: > logger="UnhandledError" Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.242460 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.243339 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.252427 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.259649 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.260343 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.260406 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.260443 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.260460 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.260474 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:28Z","lastTransitionTime":"2026-03-18T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.268644 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.276828 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.284952 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.293802 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.306145 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.316691 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.332721 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.347507 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.362715 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.362758 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.362769 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.362785 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.362796 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:28Z","lastTransitionTime":"2026-03-18T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.363112 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.464727 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.464781 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.464798 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.464827 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.464849 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:28Z","lastTransitionTime":"2026-03-18T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.539480 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.539604 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.539653 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.539686 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:09:29.539649016 +0000 UTC m=+87.437023666 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.539761 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.539812 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.539841 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.539861 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.539861 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.539857 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.540100 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.539977 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:29.539908963 +0000 UTC m=+87.437283603 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.540145 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.540165 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.539981 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.540200 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:29.5401613 +0000 UTC m=+87.437535980 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.540253 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:29.540226472 +0000 UTC m=+87.437601112 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:28 crc kubenswrapper[4983]: E0318 00:09:28.540320 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:29.540291874 +0000 UTC m=+87.437666524 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.568061 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.568115 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.568135 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.568167 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.568193 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:28Z","lastTransitionTime":"2026-03-18T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.670878 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.670975 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.671001 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.671030 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.671049 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:28Z","lastTransitionTime":"2026-03-18T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.774105 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.774267 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.774286 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.774309 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.774327 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:28Z","lastTransitionTime":"2026-03-18T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.857103 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.858785 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.861160 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.862480 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.864415 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.865484 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.866737 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.868774 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.870201 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.872293 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.873331 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.875700 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.876632 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.876693 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.876716 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.876747 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.876767 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:28Z","lastTransitionTime":"2026-03-18T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.877648 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.878811 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.880894 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.882096 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.884144 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.884995 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.886238 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.887750 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.888474 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.889844 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.890472 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.892113 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.892680 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.893522 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.895374 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.896152 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.897438 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.898082 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.899474 4983 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.899613 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.901837 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.903182 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.903732 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.905983 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.906947 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.908164 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.909055 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.910616 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.911290 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.912787 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.913638 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.915104 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.915788 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.917288 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.918022 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.920708 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.922080 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.924139 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.925218 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.927516 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.928860 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.929894 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.980083 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.980215 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.980286 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.980317 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:28 crc kubenswrapper[4983]: I0318 00:09:28.980339 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:28Z","lastTransitionTime":"2026-03-18T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.084525 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.084612 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.084639 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.084674 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.084709 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:29Z","lastTransitionTime":"2026-03-18T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.187906 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.188042 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.188072 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.188110 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.188139 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:29Z","lastTransitionTime":"2026-03-18T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.293773 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.293860 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.293879 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.293908 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.293956 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:29Z","lastTransitionTime":"2026-03-18T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.397412 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.397498 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.397518 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.397546 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.397568 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:29Z","lastTransitionTime":"2026-03-18T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.501005 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.501062 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.501082 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.501115 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.501139 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:29Z","lastTransitionTime":"2026-03-18T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.549412 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.549549 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.549596 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.549670 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:09:31.549615603 +0000 UTC m=+89.446990273 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.549770 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.549787 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.549790 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.549811 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.549891 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.549917 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.549886 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:31.549843209 +0000 UTC m=+89.447217829 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.550397 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:31.550359374 +0000 UTC m=+89.447734164 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.550423 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.550459 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.550462 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:31.550441496 +0000 UTC m=+89.447816366 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.550487 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.550630 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:31.550596481 +0000 UTC m=+89.447971191 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.550700 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.604231 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.604309 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.604327 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.604353 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.604372 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:29Z","lastTransitionTime":"2026-03-18T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.706686 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.707128 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.707286 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.707477 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.707608 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:29Z","lastTransitionTime":"2026-03-18T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.810292 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.810357 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.810382 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.810411 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.810434 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:29Z","lastTransitionTime":"2026-03-18T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.844488 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.844637 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.844705 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.844882 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.845100 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.845397 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.907016 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.907083 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.907102 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.907128 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.907154 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:29Z","lastTransitionTime":"2026-03-18T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.923173 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.928207 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.928276 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.928300 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.928327 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.928349 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:29Z","lastTransitionTime":"2026-03-18T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.945718 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.951387 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.951466 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.951488 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.951515 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.951534 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:29Z","lastTransitionTime":"2026-03-18T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.967302 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.971896 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.971978 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.971997 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.972022 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.972039 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:29Z","lastTransitionTime":"2026-03-18T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:29 crc kubenswrapper[4983]: E0318 00:09:29.987293 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.992886 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.992963 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.992983 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.993006 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:29 crc kubenswrapper[4983]: I0318 00:09:29.993026 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:29Z","lastTransitionTime":"2026-03-18T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:30 crc kubenswrapper[4983]: E0318 00:09:30.007220 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:30 crc kubenswrapper[4983]: E0318 00:09:30.007433 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.009280 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.009339 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.009357 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.009383 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.009401 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:30Z","lastTransitionTime":"2026-03-18T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.113190 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.113269 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.113293 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.113325 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.113350 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:30Z","lastTransitionTime":"2026-03-18T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.216289 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.216345 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.216362 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.216386 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.216403 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:30Z","lastTransitionTime":"2026-03-18T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.320089 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.320146 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.320167 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.320196 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.320218 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:30Z","lastTransitionTime":"2026-03-18T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.423079 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.423144 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.423167 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.423197 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.423220 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:30Z","lastTransitionTime":"2026-03-18T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.526507 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.526640 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.526660 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.526688 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.526708 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:30Z","lastTransitionTime":"2026-03-18T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.630142 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.630229 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.630256 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.630286 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.630310 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:30Z","lastTransitionTime":"2026-03-18T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.733691 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.733752 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.733775 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.733799 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.733817 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:30Z","lastTransitionTime":"2026-03-18T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.837335 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.837425 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.837441 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.837467 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.837485 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:30Z","lastTransitionTime":"2026-03-18T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.940091 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.940177 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.940197 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.940229 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:30 crc kubenswrapper[4983]: I0318 00:09:30.940246 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:30Z","lastTransitionTime":"2026-03-18T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.043339 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.043390 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.043401 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.043418 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.043432 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:31Z","lastTransitionTime":"2026-03-18T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.146382 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.146464 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.146486 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.146516 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.146539 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:31Z","lastTransitionTime":"2026-03-18T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.249145 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.249214 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.249232 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.249256 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.249275 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:31Z","lastTransitionTime":"2026-03-18T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.351495 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.351541 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.351553 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.351570 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.351582 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:31Z","lastTransitionTime":"2026-03-18T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.453251 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.453317 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.453335 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.453357 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.453375 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:31Z","lastTransitionTime":"2026-03-18T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.556114 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.556192 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.556225 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.556253 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.556273 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:31Z","lastTransitionTime":"2026-03-18T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.572800 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.572890 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.572958 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.573014 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.573035 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:09:35.572998198 +0000 UTC m=+93.470372858 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.573051 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.573086 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.573193 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.573211 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.573232 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.573253 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.573266 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.573278 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.573286 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.573238 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:35.573110212 +0000 UTC m=+93.470484822 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.573315 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:35.573303527 +0000 UTC m=+93.470678147 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.573348 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:35.573324878 +0000 UTC m=+93.470699528 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.573382 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:35.573366289 +0000 UTC m=+93.470740939 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.659418 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.659480 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.659508 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.659538 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.659558 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:31Z","lastTransitionTime":"2026-03-18T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.762075 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.762134 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.762150 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.762174 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.762242 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:31Z","lastTransitionTime":"2026-03-18T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.843866 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.844002 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.843867 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.844114 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.844582 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:31 crc kubenswrapper[4983]: E0318 00:09:31.845030 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.865622 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.865859 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.866114 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.866270 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.866411 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:31Z","lastTransitionTime":"2026-03-18T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.968309 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.968349 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.968358 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.968372 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:31 crc kubenswrapper[4983]: I0318 00:09:31.968383 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:31Z","lastTransitionTime":"2026-03-18T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.070882 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.071196 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.071307 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.071456 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.071483 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:32Z","lastTransitionTime":"2026-03-18T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.173799 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.173904 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.173922 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.173974 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.173993 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:32Z","lastTransitionTime":"2026-03-18T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.190366 4983 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.276622 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.276663 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.276673 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.276690 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.276703 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:32Z","lastTransitionTime":"2026-03-18T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.379967 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.380015 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.380031 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.380053 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.380072 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:32Z","lastTransitionTime":"2026-03-18T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.482540 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.482654 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.482673 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.482695 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.482711 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:32Z","lastTransitionTime":"2026-03-18T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.584544 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.584599 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.584621 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.584647 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.584669 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:32Z","lastTransitionTime":"2026-03-18T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.687637 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.687691 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.687708 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.687737 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.687759 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:32Z","lastTransitionTime":"2026-03-18T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.790352 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.790419 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.790442 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.790467 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.790482 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:32Z","lastTransitionTime":"2026-03-18T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.861620 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.875550 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.890357 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.892659 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.892717 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.892734 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.892757 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.892775 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:32Z","lastTransitionTime":"2026-03-18T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.903503 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.918034 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.936482 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.994567 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.994639 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.994652 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.994670 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:32 crc kubenswrapper[4983]: I0318 00:09:32.994686 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:32Z","lastTransitionTime":"2026-03-18T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.097904 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.098003 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.098029 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.098057 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.098085 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:33Z","lastTransitionTime":"2026-03-18T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.201327 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.201376 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.201393 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.201419 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.201435 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:33Z","lastTransitionTime":"2026-03-18T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.303862 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.303999 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.304022 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.304051 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.304073 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:33Z","lastTransitionTime":"2026-03-18T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.406466 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.406545 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.406571 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.406602 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.406627 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:33Z","lastTransitionTime":"2026-03-18T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.509541 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.509613 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.509639 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.509668 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.509689 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:33Z","lastTransitionTime":"2026-03-18T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.612054 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.612310 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.612373 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.612447 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.612529 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:33Z","lastTransitionTime":"2026-03-18T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.714997 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.715313 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.715459 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.715597 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.715757 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:33Z","lastTransitionTime":"2026-03-18T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.818195 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.818276 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.818301 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.818330 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.818353 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:33Z","lastTransitionTime":"2026-03-18T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.843859 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:33 crc kubenswrapper[4983]: E0318 00:09:33.844249 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.844033 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.843869 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:33 crc kubenswrapper[4983]: E0318 00:09:33.844654 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:33 crc kubenswrapper[4983]: E0318 00:09:33.844789 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.920433 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.920687 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.920772 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.920861 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:33 crc kubenswrapper[4983]: I0318 00:09:33.920970 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:33Z","lastTransitionTime":"2026-03-18T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.023627 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.023671 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.023685 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.023707 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.023717 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:34Z","lastTransitionTime":"2026-03-18T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.126264 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.126299 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.126307 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.126319 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.126328 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:34Z","lastTransitionTime":"2026-03-18T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.229025 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.229072 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.229087 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.229110 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.229125 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:34Z","lastTransitionTime":"2026-03-18T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.331999 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.332684 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.332959 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.333183 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.333383 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:34Z","lastTransitionTime":"2026-03-18T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.436829 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.436909 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.436962 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.436988 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.437004 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:34Z","lastTransitionTime":"2026-03-18T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.539427 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.539461 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.539471 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.539485 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.539496 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:34Z","lastTransitionTime":"2026-03-18T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.642123 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.642377 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.642465 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.642557 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.642620 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:34Z","lastTransitionTime":"2026-03-18T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.745399 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.745450 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.745458 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.745470 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.745479 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:34Z","lastTransitionTime":"2026-03-18T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.764485 4983 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.848664 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.848701 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.848714 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.848729 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.848743 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:34Z","lastTransitionTime":"2026-03-18T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.951305 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.951365 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.951387 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.951414 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:34 crc kubenswrapper[4983]: I0318 00:09:34.951434 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:34Z","lastTransitionTime":"2026-03-18T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.054070 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.054138 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.054163 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.054193 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.054214 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:35Z","lastTransitionTime":"2026-03-18T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.157105 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.157162 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.157178 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.157201 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.157219 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:35Z","lastTransitionTime":"2026-03-18T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.259425 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.259478 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.259500 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.259528 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.259547 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:35Z","lastTransitionTime":"2026-03-18T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.361752 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.361838 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.361868 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.361901 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.361958 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:35Z","lastTransitionTime":"2026-03-18T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.464354 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.464438 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.464457 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.464486 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.464508 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:35Z","lastTransitionTime":"2026-03-18T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.567740 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.567798 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.567817 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.567840 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.567859 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:35Z","lastTransitionTime":"2026-03-18T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.607898 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.608036 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.608086 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:09:43.608052392 +0000 UTC m=+101.505427032 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.608138 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.608150 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.608198 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:43.608180316 +0000 UTC m=+101.505554956 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.608226 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.608263 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.608325 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.608356 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.608378 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.608442 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:43.608420943 +0000 UTC m=+101.505795593 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.608380 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.608483 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.608498 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.608545 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:43.608529686 +0000 UTC m=+101.505904336 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.608648 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.608695 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:43.60868228 +0000 UTC m=+101.506056920 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.669515 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.669561 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.669573 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.669592 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.669603 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:35Z","lastTransitionTime":"2026-03-18T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.772198 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.772253 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.772271 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.772296 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.772313 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:35Z","lastTransitionTime":"2026-03-18T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.843853 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.843887 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.844195 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.844470 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.844680 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:35 crc kubenswrapper[4983]: E0318 00:09:35.844637 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.874766 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.875165 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.875469 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.875773 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.876084 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:35Z","lastTransitionTime":"2026-03-18T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.978970 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.979022 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.979041 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.979065 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:35 crc kubenswrapper[4983]: I0318 00:09:35.979084 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:35Z","lastTransitionTime":"2026-03-18T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.081583 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.081638 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.081655 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.081678 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.081696 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:36Z","lastTransitionTime":"2026-03-18T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.184747 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.184806 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.184825 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.184852 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.184871 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:36Z","lastTransitionTime":"2026-03-18T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.287512 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.287568 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.287587 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.287611 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.287627 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:36Z","lastTransitionTime":"2026-03-18T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.390543 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.390606 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.390623 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.390648 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.390666 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:36Z","lastTransitionTime":"2026-03-18T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.493446 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.493481 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.493491 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.493507 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.493517 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:36Z","lastTransitionTime":"2026-03-18T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.596554 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.596857 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.597040 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.597328 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.597511 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:36Z","lastTransitionTime":"2026-03-18T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.701455 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.701542 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.701591 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.701616 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.701636 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:36Z","lastTransitionTime":"2026-03-18T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.804266 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.804350 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.804394 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.804427 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.804452 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:36Z","lastTransitionTime":"2026-03-18T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.907497 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.907577 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.907595 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.907620 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:36 crc kubenswrapper[4983]: I0318 00:09:36.907639 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:36Z","lastTransitionTime":"2026-03-18T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.011188 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.011230 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.011241 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.011256 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.011269 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:37Z","lastTransitionTime":"2026-03-18T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.114103 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.114150 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.114161 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.114180 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.114192 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:37Z","lastTransitionTime":"2026-03-18T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.217777 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.217830 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.217848 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.217874 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.217890 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:37Z","lastTransitionTime":"2026-03-18T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.320999 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.321054 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.321076 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.321101 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.321118 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:37Z","lastTransitionTime":"2026-03-18T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.423806 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.423860 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.423880 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.423910 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.423953 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:37Z","lastTransitionTime":"2026-03-18T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.527545 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.527622 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.527640 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.527666 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.527684 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:37Z","lastTransitionTime":"2026-03-18T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.630009 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.630054 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.630072 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.630094 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.630111 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:37Z","lastTransitionTime":"2026-03-18T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.732907 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.733011 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.733032 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.733059 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.733081 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:37Z","lastTransitionTime":"2026-03-18T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.836254 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.836308 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.836318 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.836336 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.836347 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:37Z","lastTransitionTime":"2026-03-18T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.843800 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.843872 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.843894 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:37 crc kubenswrapper[4983]: E0318 00:09:37.844052 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:37 crc kubenswrapper[4983]: E0318 00:09:37.844122 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:37 crc kubenswrapper[4983]: E0318 00:09:37.844269 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.939049 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.939101 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.939115 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.939134 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:37 crc kubenswrapper[4983]: I0318 00:09:37.939147 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:37Z","lastTransitionTime":"2026-03-18T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.042656 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.042733 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.042750 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.042771 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.042780 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:38Z","lastTransitionTime":"2026-03-18T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.145815 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.145871 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.145887 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.145904 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.145914 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:38Z","lastTransitionTime":"2026-03-18T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.249062 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.249124 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.249137 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.249152 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.249162 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:38Z","lastTransitionTime":"2026-03-18T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.352079 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.352134 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.352153 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.352175 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.352192 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:38Z","lastTransitionTime":"2026-03-18T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.460264 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.461580 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.461874 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.462190 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.462374 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:38Z","lastTransitionTime":"2026-03-18T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.566595 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.566684 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.566710 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.566746 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.566774 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:38Z","lastTransitionTime":"2026-03-18T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.669343 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.669390 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.669402 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.669420 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.669434 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:38Z","lastTransitionTime":"2026-03-18T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.772564 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.772632 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.772655 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.772692 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.772716 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:38Z","lastTransitionTime":"2026-03-18T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.861457 4983 scope.go:117] "RemoveContainer" containerID="b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626" Mar 18 00:09:38 crc kubenswrapper[4983]: E0318 00:09:38.861756 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.863464 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.865280 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.876070 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.876112 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.876122 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.876140 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.876151 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:38Z","lastTransitionTime":"2026-03-18T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.982380 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.982671 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.983239 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.983799 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:38 crc kubenswrapper[4983]: I0318 00:09:38.984218 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:38Z","lastTransitionTime":"2026-03-18T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.088761 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.089231 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.089242 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.089259 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.089271 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:39Z","lastTransitionTime":"2026-03-18T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.193035 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.193126 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.193149 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.193180 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.193202 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:39Z","lastTransitionTime":"2026-03-18T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.271051 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28"} Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.271122 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a"} Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.272362 4983 scope.go:117] "RemoveContainer" containerID="b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626" Mar 18 00:09:39 crc kubenswrapper[4983]: E0318 00:09:39.272679 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.290448 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.295739 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.295814 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.295845 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.295879 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.295905 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:39Z","lastTransitionTime":"2026-03-18T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.307709 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.337859 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.349851 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.363184 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.374714 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.388028 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.400269 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.400334 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.400356 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.400384 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.400404 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:39Z","lastTransitionTime":"2026-03-18T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.407135 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.503842 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.503909 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.503950 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.503980 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.503999 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:39Z","lastTransitionTime":"2026-03-18T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.607215 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.607909 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.608312 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.608523 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.608654 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:39Z","lastTransitionTime":"2026-03-18T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.711274 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.711345 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.711358 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.711378 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.711391 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:39Z","lastTransitionTime":"2026-03-18T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.814540 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.814598 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.814619 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.814653 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.814671 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:39Z","lastTransitionTime":"2026-03-18T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.843538 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.843583 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.843676 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:39 crc kubenswrapper[4983]: E0318 00:09:39.843823 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:39 crc kubenswrapper[4983]: E0318 00:09:39.843992 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:39 crc kubenswrapper[4983]: E0318 00:09:39.844155 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.918653 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.918719 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.918735 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.918760 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:39 crc kubenswrapper[4983]: I0318 00:09:39.918777 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:39Z","lastTransitionTime":"2026-03-18T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.021995 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.022078 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.022097 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.022123 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.022141 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.035612 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.035672 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.035691 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.035715 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.035736 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: E0318 00:09:40.055988 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.061998 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.062036 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.062046 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.062063 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.062073 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: E0318 00:09:40.078500 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.083577 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.083637 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.083654 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.083680 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.083697 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: E0318 00:09:40.101568 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.107087 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.107187 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.107260 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.107331 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.107396 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: E0318 00:09:40.127640 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.133820 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.133875 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.133891 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.133909 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.133923 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: E0318 00:09:40.152212 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:40 crc kubenswrapper[4983]: E0318 00:09:40.152521 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.154379 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.154410 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.154426 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.154444 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.154458 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.257322 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.257651 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.257860 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.258178 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.258329 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.361021 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.361115 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.361134 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.361160 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.361177 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.463567 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.463626 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.463647 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.463670 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.463689 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.566731 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.566771 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.566781 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.566797 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.566807 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.670018 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.670096 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.670119 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.670148 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.670171 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.772790 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.773065 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.773167 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.773266 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.773357 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.875749 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.875813 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.875838 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.875868 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.875887 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.979119 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.979216 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.979245 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.979274 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:40 crc kubenswrapper[4983]: I0318 00:09:40.979296 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:40Z","lastTransitionTime":"2026-03-18T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.081963 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.082036 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.082059 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.082088 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.082110 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:41Z","lastTransitionTime":"2026-03-18T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.185842 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.185910 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.185953 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.185977 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.185995 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:41Z","lastTransitionTime":"2026-03-18T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.288844 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.288897 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.288913 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.288962 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.288982 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:41Z","lastTransitionTime":"2026-03-18T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.392031 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.392131 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.392151 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.392211 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.392231 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:41Z","lastTransitionTime":"2026-03-18T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.495200 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.495259 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.495277 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.495300 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.495324 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:41Z","lastTransitionTime":"2026-03-18T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.598463 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.599283 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.599486 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.599635 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.599779 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:41Z","lastTransitionTime":"2026-03-18T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.702986 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.703069 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.703089 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.703114 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.703131 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:41Z","lastTransitionTime":"2026-03-18T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.806696 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.807095 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.807275 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.807604 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.807804 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:41Z","lastTransitionTime":"2026-03-18T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.843785 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:41 crc kubenswrapper[4983]: E0318 00:09:41.844151 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.843869 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:41 crc kubenswrapper[4983]: E0318 00:09:41.844594 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.843863 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:41 crc kubenswrapper[4983]: E0318 00:09:41.845058 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.910631 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.910921 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.911162 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.911306 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:41 crc kubenswrapper[4983]: I0318 00:09:41.911447 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:41Z","lastTransitionTime":"2026-03-18T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.014759 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.014815 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.014831 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.014854 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.014872 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:42Z","lastTransitionTime":"2026-03-18T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.171189 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.171254 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.171272 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.171295 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.171315 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:42Z","lastTransitionTime":"2026-03-18T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.276742 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.276896 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.276993 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.277109 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.277150 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:42Z","lastTransitionTime":"2026-03-18T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.380775 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.380831 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.380848 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.380872 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.380889 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:42Z","lastTransitionTime":"2026-03-18T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.484424 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.484532 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.484551 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.484578 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.484596 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:42Z","lastTransitionTime":"2026-03-18T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.523251 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-cvs9w"] Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.523668 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-cvs9w" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.527199 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.527525 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.528036 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.555126 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.575218 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.587833 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.587894 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.587919 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.587980 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.588004 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:42Z","lastTransitionTime":"2026-03-18T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.590845 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.628595 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.645142 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.663020 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.674189 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ca9fe29e-1efb-4f9c-b0e4-ca34025d56de-hosts-file\") pod \"node-resolver-cvs9w\" (UID: \"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\") " pod="openshift-dns/node-resolver-cvs9w" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.674273 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw7t4\" (UniqueName: \"kubernetes.io/projected/ca9fe29e-1efb-4f9c-b0e4-ca34025d56de-kube-api-access-gw7t4\") pod \"node-resolver-cvs9w\" (UID: \"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\") " pod="openshift-dns/node-resolver-cvs9w" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.682045 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.690103 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.690148 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.690166 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.690191 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.690209 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:42Z","lastTransitionTime":"2026-03-18T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.703073 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.724719 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.775064 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ca9fe29e-1efb-4f9c-b0e4-ca34025d56de-hosts-file\") pod \"node-resolver-cvs9w\" (UID: \"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\") " pod="openshift-dns/node-resolver-cvs9w" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.775126 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw7t4\" (UniqueName: \"kubernetes.io/projected/ca9fe29e-1efb-4f9c-b0e4-ca34025d56de-kube-api-access-gw7t4\") pod \"node-resolver-cvs9w\" (UID: \"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\") " pod="openshift-dns/node-resolver-cvs9w" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.775304 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ca9fe29e-1efb-4f9c-b0e4-ca34025d56de-hosts-file\") pod \"node-resolver-cvs9w\" (UID: \"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\") " pod="openshift-dns/node-resolver-cvs9w" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.794106 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.794208 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.794226 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.794253 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.794270 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:42Z","lastTransitionTime":"2026-03-18T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.815290 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw7t4\" (UniqueName: \"kubernetes.io/projected/ca9fe29e-1efb-4f9c-b0e4-ca34025d56de-kube-api-access-gw7t4\") pod \"node-resolver-cvs9w\" (UID: \"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\") " pod="openshift-dns/node-resolver-cvs9w" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.849199 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-cvs9w" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.866920 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: W0318 00:09:42.872157 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca9fe29e_1efb_4f9c_b0e4_ca34025d56de.slice/crio-7d8b4d377f1bd42c696e82cff94fb63779d35beb18c8fcaafbd661ee5dfc6c2f WatchSource:0}: Error finding container 7d8b4d377f1bd42c696e82cff94fb63779d35beb18c8fcaafbd661ee5dfc6c2f: Status 404 returned error can't find the container with id 7d8b4d377f1bd42c696e82cff94fb63779d35beb18c8fcaafbd661ee5dfc6c2f Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.886884 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.900539 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.900611 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.900634 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.900669 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.900691 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:42Z","lastTransitionTime":"2026-03-18T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.902349 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.930845 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-jczfs"] Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.931553 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.931486 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.935419 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.936316 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-6nblt"] Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.937048 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-69n75"] Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.937918 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.938516 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6nblt" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.939194 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.939394 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.939753 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.940610 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.941670 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.942634 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.942774 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.943031 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.943234 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.945040 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.945749 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.955488 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.976679 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:42 crc kubenswrapper[4983]: I0318 00:09:42.997483 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.004163 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.004192 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.004200 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.004213 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.004222 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:43Z","lastTransitionTime":"2026-03-18T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.016098 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.039298 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.054537 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.070451 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077413 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-run-multus-certs\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077462 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-system-cni-dir\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077483 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-multus-conf-dir\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077505 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/482e2afe-48be-4ea2-8bda-52849c901b52-rootfs\") pod \"machine-config-daemon-jczfs\" (UID: \"482e2afe-48be-4ea2-8bda-52849c901b52\") " pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077525 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-cni-binary-copy\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077549 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-system-cni-dir\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077568 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-var-lib-cni-bin\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077589 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-run-k8s-cni-cncf-io\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077609 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0853f8de-9bfc-4053-acd0-8893122d2d62-cni-binary-copy\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077626 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-cnibin\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077653 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-var-lib-cni-multus\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077674 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vmw7\" (UniqueName: \"kubernetes.io/projected/0853f8de-9bfc-4053-acd0-8893122d2d62-kube-api-access-4vmw7\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077703 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-etc-kubernetes\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077725 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-run-netns\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077744 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/482e2afe-48be-4ea2-8bda-52849c901b52-proxy-tls\") pod \"machine-config-daemon-jczfs\" (UID: \"482e2afe-48be-4ea2-8bda-52849c901b52\") " pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077766 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfjxb\" (UniqueName: \"kubernetes.io/projected/482e2afe-48be-4ea2-8bda-52849c901b52-kube-api-access-kfjxb\") pod \"machine-config-daemon-jczfs\" (UID: \"482e2afe-48be-4ea2-8bda-52849c901b52\") " pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077788 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-cnibin\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077808 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077830 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-os-release\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077855 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-multus-socket-dir-parent\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077875 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8459\" (UniqueName: \"kubernetes.io/projected/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-kube-api-access-g8459\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077895 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-hostroot\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077915 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/482e2afe-48be-4ea2-8bda-52849c901b52-mcd-auth-proxy-config\") pod \"machine-config-daemon-jczfs\" (UID: \"482e2afe-48be-4ea2-8bda-52849c901b52\") " pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077963 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0853f8de-9bfc-4053-acd0-8893122d2d62-multus-daemon-config\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.077996 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.078016 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-multus-cni-dir\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.078036 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-var-lib-kubelet\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.078064 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-os-release\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.083227 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.097093 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.105902 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.105954 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.105965 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.105981 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.105992 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:43Z","lastTransitionTime":"2026-03-18T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.114549 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.131089 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.150415 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.165557 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179347 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-run-multus-certs\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179384 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-system-cni-dir\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179405 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-multus-conf-dir\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179428 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/482e2afe-48be-4ea2-8bda-52849c901b52-rootfs\") pod \"machine-config-daemon-jczfs\" (UID: \"482e2afe-48be-4ea2-8bda-52849c901b52\") " pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179448 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-cni-binary-copy\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179495 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-system-cni-dir\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179515 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-var-lib-cni-bin\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179535 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-run-k8s-cni-cncf-io\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179557 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0853f8de-9bfc-4053-acd0-8893122d2d62-cni-binary-copy\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179576 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-cnibin\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179580 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/482e2afe-48be-4ea2-8bda-52849c901b52-rootfs\") pod \"machine-config-daemon-jczfs\" (UID: \"482e2afe-48be-4ea2-8bda-52849c901b52\") " pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179596 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-var-lib-cni-multus\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179617 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vmw7\" (UniqueName: \"kubernetes.io/projected/0853f8de-9bfc-4053-acd0-8893122d2d62-kube-api-access-4vmw7\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179650 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-system-cni-dir\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179650 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-etc-kubernetes\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179670 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-multus-conf-dir\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179678 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-run-netns\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179706 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/482e2afe-48be-4ea2-8bda-52849c901b52-proxy-tls\") pod \"machine-config-daemon-jczfs\" (UID: \"482e2afe-48be-4ea2-8bda-52849c901b52\") " pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179735 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfjxb\" (UniqueName: \"kubernetes.io/projected/482e2afe-48be-4ea2-8bda-52849c901b52-kube-api-access-kfjxb\") pod \"machine-config-daemon-jczfs\" (UID: \"482e2afe-48be-4ea2-8bda-52849c901b52\") " pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179761 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-cnibin\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179788 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179816 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-os-release\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179845 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-multus-socket-dir-parent\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179875 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8459\" (UniqueName: \"kubernetes.io/projected/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-kube-api-access-g8459\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179899 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-hostroot\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179943 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/482e2afe-48be-4ea2-8bda-52849c901b52-mcd-auth-proxy-config\") pod \"machine-config-daemon-jczfs\" (UID: \"482e2afe-48be-4ea2-8bda-52849c901b52\") " pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179967 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0853f8de-9bfc-4053-acd0-8893122d2d62-multus-daemon-config\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179997 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.180017 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-multus-cni-dir\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.180036 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-var-lib-kubelet\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.180067 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-os-release\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.180149 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-os-release\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.180205 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-cnibin\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.180238 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-var-lib-cni-multus\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.180376 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0853f8de-9bfc-4053-acd0-8893122d2d62-cni-binary-copy\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.179620 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-run-multus-certs\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.180564 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-etc-kubernetes\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.180607 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-run-netns\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.181015 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-cni-binary-copy\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.181064 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-system-cni-dir\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.181093 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-var-lib-cni-bin\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.181121 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-run-k8s-cni-cncf-io\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.181292 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.181742 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-os-release\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.181990 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-host-var-lib-kubelet\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.182027 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-cnibin\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.182075 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-multus-cni-dir\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.182111 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-multus-socket-dir-parent\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.182148 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0853f8de-9bfc-4053-acd0-8893122d2d62-hostroot\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.182206 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.182276 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.183263 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/482e2afe-48be-4ea2-8bda-52849c901b52-mcd-auth-proxy-config\") pod \"machine-config-daemon-jczfs\" (UID: \"482e2afe-48be-4ea2-8bda-52849c901b52\") " pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.183879 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0853f8de-9bfc-4053-acd0-8893122d2d62-multus-daemon-config\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.185767 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/482e2afe-48be-4ea2-8bda-52849c901b52-proxy-tls\") pod \"machine-config-daemon-jczfs\" (UID: \"482e2afe-48be-4ea2-8bda-52849c901b52\") " pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.202675 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8459\" (UniqueName: \"kubernetes.io/projected/72a1209f-2f3a-48e3-ade2-7aa5824f45eb-kube-api-access-g8459\") pod \"multus-additional-cni-plugins-69n75\" (UID: \"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\") " pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.203281 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfjxb\" (UniqueName: \"kubernetes.io/projected/482e2afe-48be-4ea2-8bda-52849c901b52-kube-api-access-kfjxb\") pod \"machine-config-daemon-jczfs\" (UID: \"482e2afe-48be-4ea2-8bda-52849c901b52\") " pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.204074 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.207640 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vmw7\" (UniqueName: \"kubernetes.io/projected/0853f8de-9bfc-4053-acd0-8893122d2d62-kube-api-access-4vmw7\") pod \"multus-6nblt\" (UID: \"0853f8de-9bfc-4053-acd0-8893122d2d62\") " pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.208877 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.208908 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.208921 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.208959 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.208971 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:43Z","lastTransitionTime":"2026-03-18T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.219064 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.235998 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.258610 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:09:43 crc kubenswrapper[4983]: W0318 00:09:43.271385 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod482e2afe_48be_4ea2_8bda_52849c901b52.slice/crio-fc18ea92f804ffc7f80d2e70828ab02e23a463cc9053c48d31a1dbcd8049702a WatchSource:0}: Error finding container fc18ea92f804ffc7f80d2e70828ab02e23a463cc9053c48d31a1dbcd8049702a: Status 404 returned error can't find the container with id fc18ea92f804ffc7f80d2e70828ab02e23a463cc9053c48d31a1dbcd8049702a Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.273453 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-69n75" Mar 18 00:09:43 crc kubenswrapper[4983]: W0318 00:09:43.283125 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72a1209f_2f3a_48e3_ade2_7aa5824f45eb.slice/crio-d332c5a3fe302b5fe1cdf94836e3724b0454016d79eb19c1cfb1725764817176 WatchSource:0}: Error finding container d332c5a3fe302b5fe1cdf94836e3724b0454016d79eb19c1cfb1725764817176: Status 404 returned error can't find the container with id d332c5a3fe302b5fe1cdf94836e3724b0454016d79eb19c1cfb1725764817176 Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.283381 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerStarted","Data":"fc18ea92f804ffc7f80d2e70828ab02e23a463cc9053c48d31a1dbcd8049702a"} Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.284763 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722"} Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.286181 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6nblt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.286310 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-cvs9w" event={"ID":"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de","Type":"ContainerStarted","Data":"fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c"} Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.286362 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-cvs9w" event={"ID":"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de","Type":"ContainerStarted","Data":"7d8b4d377f1bd42c696e82cff94fb63779d35beb18c8fcaafbd661ee5dfc6c2f"} Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.303644 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.312322 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.312368 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.312378 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.312395 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.312405 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:43Z","lastTransitionTime":"2026-03-18T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:43 crc kubenswrapper[4983]: W0318 00:09:43.312692 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0853f8de_9bfc_4053_acd0_8893122d2d62.slice/crio-e1296456c0c8908c285323a36d55273577760a04076a23b102c25849f3df40a5 WatchSource:0}: Error finding container e1296456c0c8908c285323a36d55273577760a04076a23b102c25849f3df40a5: Status 404 returned error can't find the container with id e1296456c0c8908c285323a36d55273577760a04076a23b102c25849f3df40a5 Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.318174 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.323765 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4r8sk"] Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.325441 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.330133 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.332053 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.332127 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.332468 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.332513 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.332547 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.332773 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.337047 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.354685 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.370416 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.394356 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.415437 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.417444 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.417489 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.417507 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.417670 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.417691 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:43Z","lastTransitionTime":"2026-03-18T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.427879 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.441771 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.453529 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.469384 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.481768 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-var-lib-openvswitch\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.481813 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-run-netns\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.481838 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-ovnkube-config\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.481860 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-cni-netd\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.481882 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-node-log\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.481986 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-systemd\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482038 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-log-socket\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482093 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482193 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-openvswitch\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482281 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-ovnkube-script-lib\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482367 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-kubelet\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482404 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-env-overrides\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482436 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-cni-bin\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482472 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/938043fe-450c-4477-8f55-312a843b0706-ovn-node-metrics-cert\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482509 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-etc-openvswitch\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482536 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-systemd-units\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482562 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-slash\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482587 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-run-ovn-kubernetes\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482615 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-ovn\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.482667 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tj4g\" (UniqueName: \"kubernetes.io/projected/938043fe-450c-4477-8f55-312a843b0706-kube-api-access-6tj4g\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.486257 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.501301 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.513427 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.519973 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.520020 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.520038 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.520060 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.520078 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:43Z","lastTransitionTime":"2026-03-18T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.532518 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.550172 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.567529 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.582902 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583300 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-etc-openvswitch\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583362 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-run-ovn-kubernetes\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583396 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-systemd-units\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583428 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-slash\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583469 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-etc-openvswitch\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583506 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-run-ovn-kubernetes\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583541 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-ovn\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583486 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-ovn\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583552 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-systemd-units\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583603 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tj4g\" (UniqueName: \"kubernetes.io/projected/938043fe-450c-4477-8f55-312a843b0706-kube-api-access-6tj4g\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583556 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-slash\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583681 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-var-lib-openvswitch\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583726 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-ovnkube-config\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583762 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-run-netns\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583799 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-var-lib-openvswitch\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583807 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-cni-netd\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583845 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-cni-netd\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583868 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-systemd\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583895 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-run-netns\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583900 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-node-log\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583959 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-openvswitch\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583976 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-node-log\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.583979 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-log-socket\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.584019 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-openvswitch\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.584029 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.584049 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-systemd\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.584000 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-log-socket\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.584064 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-ovnkube-script-lib\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.584086 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.584166 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-kubelet\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.584223 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-env-overrides\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.584283 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-cni-bin\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.584316 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/938043fe-450c-4477-8f55-312a843b0706-ovn-node-metrics-cert\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.584973 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-ovnkube-script-lib\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.585074 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-ovnkube-config\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.585152 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-kubelet\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.585170 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-cni-bin\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.585383 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-env-overrides\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.590638 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/938043fe-450c-4477-8f55-312a843b0706-ovn-node-metrics-cert\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.623306 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.623351 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.623363 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.623382 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.623393 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:43Z","lastTransitionTime":"2026-03-18T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.626676 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.630883 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tj4g\" (UniqueName: \"kubernetes.io/projected/938043fe-450c-4477-8f55-312a843b0706-kube-api-access-6tj4g\") pod \"ovnkube-node-4r8sk\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.644026 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.654556 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.654812 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: W0318 00:09:43.664947 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod938043fe_450c_4477_8f55_312a843b0706.slice/crio-b48225f4b6d13dc79a0cb55a03dd7dfcf45305c9e6d68df17c10c35e9b247c5e WatchSource:0}: Error finding container b48225f4b6d13dc79a0cb55a03dd7dfcf45305c9e6d68df17c10c35e9b247c5e: Status 404 returned error can't find the container with id b48225f4b6d13dc79a0cb55a03dd7dfcf45305c9e6d68df17c10c35e9b247c5e Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.678423 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.685162 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.685249 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.685289 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.685316 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.685393 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:09:59.685365226 +0000 UTC m=+117.582739846 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.685424 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.685443 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.685455 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.685480 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.685511 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.685482 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.685500 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:59.685485059 +0000 UTC m=+117.582859679 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.685586 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.685612 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.685626 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.685626 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:59.685596752 +0000 UTC m=+117.582971372 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.685657 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:59.685647914 +0000 UTC m=+117.583022534 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.685673 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:59.685664974 +0000 UTC m=+117.583039594 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.703968 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.716620 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.725304 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.725361 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.725373 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.725389 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.725402 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:43Z","lastTransitionTime":"2026-03-18T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.733031 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.827746 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.827806 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.827817 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.827832 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.827843 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:43Z","lastTransitionTime":"2026-03-18T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.844324 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.844459 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.844506 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.844655 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.845068 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:43 crc kubenswrapper[4983]: E0318 00:09:43.845183 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.931194 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.931270 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.931293 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.931323 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:43 crc kubenswrapper[4983]: I0318 00:09:43.931344 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:43Z","lastTransitionTime":"2026-03-18T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.034452 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.034520 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.034536 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.034561 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.034577 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:44Z","lastTransitionTime":"2026-03-18T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.137432 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.137486 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.137503 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.137525 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.137542 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:44Z","lastTransitionTime":"2026-03-18T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.240189 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.240257 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.240278 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.240308 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.240329 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:44Z","lastTransitionTime":"2026-03-18T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.290971 4983 generic.go:334] "Generic (PLEG): container finished" podID="72a1209f-2f3a-48e3-ade2-7aa5824f45eb" containerID="5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7" exitCode=0 Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.291043 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" event={"ID":"72a1209f-2f3a-48e3-ade2-7aa5824f45eb","Type":"ContainerDied","Data":"5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.291069 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" event={"ID":"72a1209f-2f3a-48e3-ade2-7aa5824f45eb","Type":"ContainerStarted","Data":"d332c5a3fe302b5fe1cdf94836e3724b0454016d79eb19c1cfb1725764817176"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.293988 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerStarted","Data":"c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.294008 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerStarted","Data":"ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.295549 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.297851 4983 generic.go:334] "Generic (PLEG): container finished" podID="938043fe-450c-4477-8f55-312a843b0706" containerID="5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9" exitCode=0 Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.297972 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.298028 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerStarted","Data":"b48225f4b6d13dc79a0cb55a03dd7dfcf45305c9e6d68df17c10c35e9b247c5e"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.300217 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6nblt" event={"ID":"0853f8de-9bfc-4053-acd0-8893122d2d62","Type":"ContainerStarted","Data":"ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.300291 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6nblt" event={"ID":"0853f8de-9bfc-4053-acd0-8893122d2d62","Type":"ContainerStarted","Data":"e1296456c0c8908c285323a36d55273577760a04076a23b102c25849f3df40a5"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.322643 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.342863 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.342901 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.342910 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.342949 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.342959 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:44Z","lastTransitionTime":"2026-03-18T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.348060 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.376162 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.396663 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.425621 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.442060 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.454841 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.454870 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.454880 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.454898 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.454909 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:44Z","lastTransitionTime":"2026-03-18T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.459399 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.472721 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.484964 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.496148 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.516059 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.527730 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.544623 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.558238 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.558278 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.558290 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.558306 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.558318 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:44Z","lastTransitionTime":"2026-03-18T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.562414 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.582221 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.590847 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.609314 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.624624 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.641489 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.656781 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.660646 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.660719 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.660735 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.660758 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.660784 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:44Z","lastTransitionTime":"2026-03-18T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.669783 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.682252 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.695655 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.710384 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.725516 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.739205 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:44Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.763042 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.763070 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.763078 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.763091 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.763099 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:44Z","lastTransitionTime":"2026-03-18T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.865471 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.865514 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.865523 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.865537 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.865547 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:44Z","lastTransitionTime":"2026-03-18T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.970527 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.970853 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.970865 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.970884 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:44 crc kubenswrapper[4983]: I0318 00:09:44.970897 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:44Z","lastTransitionTime":"2026-03-18T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.073708 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.073732 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.073739 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.073751 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.073759 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:45Z","lastTransitionTime":"2026-03-18T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.176440 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.176669 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.176683 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.176700 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.176712 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:45Z","lastTransitionTime":"2026-03-18T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.279472 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.279503 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.279517 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.279535 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.279548 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:45Z","lastTransitionTime":"2026-03-18T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.307090 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerStarted","Data":"1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.307138 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerStarted","Data":"0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.307151 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerStarted","Data":"5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.307165 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerStarted","Data":"8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.307178 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerStarted","Data":"5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.307189 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerStarted","Data":"f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.308603 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" event={"ID":"72a1209f-2f3a-48e3-ade2-7aa5824f45eb","Type":"ContainerStarted","Data":"730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.325026 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:45Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.336975 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:45Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.351798 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:45Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.370283 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:45Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.381984 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.382012 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.382021 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.382035 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.382045 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:45Z","lastTransitionTime":"2026-03-18T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.382338 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:45Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.397214 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:45Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.414071 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:45Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.441162 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:45Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.451300 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:45Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.465043 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:45Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.476193 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:45Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.484660 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.484724 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.484749 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.484783 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.484810 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:45Z","lastTransitionTime":"2026-03-18T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.488097 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:45Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.508106 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:45Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.587397 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.587452 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.587470 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.587494 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.587513 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:45Z","lastTransitionTime":"2026-03-18T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.690567 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.690645 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.690664 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.690691 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.690708 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:45Z","lastTransitionTime":"2026-03-18T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.793569 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.793626 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.793644 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.793668 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.793690 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:45Z","lastTransitionTime":"2026-03-18T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.843494 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.843593 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.843674 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:45 crc kubenswrapper[4983]: E0318 00:09:45.843665 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:45 crc kubenswrapper[4983]: E0318 00:09:45.843867 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:45 crc kubenswrapper[4983]: E0318 00:09:45.844024 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.896274 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.896320 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.896337 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.896359 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:45 crc kubenswrapper[4983]: I0318 00:09:45.896377 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:45Z","lastTransitionTime":"2026-03-18T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.000024 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.000076 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.000092 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.000117 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.000133 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:46Z","lastTransitionTime":"2026-03-18T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.102874 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.102916 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.102967 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.102991 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.103009 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:46Z","lastTransitionTime":"2026-03-18T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.206549 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.206597 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.206613 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.206637 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.206653 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:46Z","lastTransitionTime":"2026-03-18T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.310669 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.310728 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.310749 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.310778 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.310798 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:46Z","lastTransitionTime":"2026-03-18T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.320277 4983 generic.go:334] "Generic (PLEG): container finished" podID="72a1209f-2f3a-48e3-ade2-7aa5824f45eb" containerID="730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d" exitCode=0 Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.320351 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" event={"ID":"72a1209f-2f3a-48e3-ade2-7aa5824f45eb","Type":"ContainerDied","Data":"730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d"} Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.362669 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:46Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.380364 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:46Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.403206 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:46Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.413532 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.413707 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.413747 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.413779 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.413799 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:46Z","lastTransitionTime":"2026-03-18T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.423701 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:46Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.441558 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:46Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.466573 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:46Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.485636 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:46Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.504945 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:46Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.516405 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.516465 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.516488 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.516519 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.516543 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:46Z","lastTransitionTime":"2026-03-18T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.527900 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:46Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.543620 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:46Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.560261 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:46Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.579338 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:46Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.595030 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:46Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.618712 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.618748 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.618757 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.618771 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.618808 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:46Z","lastTransitionTime":"2026-03-18T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.721300 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.721339 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.721348 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.721364 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.721374 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:46Z","lastTransitionTime":"2026-03-18T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.824200 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.824249 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.824260 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.824279 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.824290 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:46Z","lastTransitionTime":"2026-03-18T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.927482 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.927530 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.927547 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.927569 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:46 crc kubenswrapper[4983]: I0318 00:09:46.927586 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:46Z","lastTransitionTime":"2026-03-18T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.031020 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.031086 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.031105 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.031131 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.031150 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:47Z","lastTransitionTime":"2026-03-18T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.134732 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.134818 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.134837 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.134863 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.134880 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:47Z","lastTransitionTime":"2026-03-18T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.237790 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.237854 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.237872 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.237901 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.237920 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:47Z","lastTransitionTime":"2026-03-18T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.326656 4983 generic.go:334] "Generic (PLEG): container finished" podID="72a1209f-2f3a-48e3-ade2-7aa5824f45eb" containerID="1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb" exitCode=0 Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.326716 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" event={"ID":"72a1209f-2f3a-48e3-ade2-7aa5824f45eb","Type":"ContainerDied","Data":"1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb"} Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.340892 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.340972 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.340990 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.341013 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.341031 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:47Z","lastTransitionTime":"2026-03-18T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.376611 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.393884 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.418631 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.442650 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.444643 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.444681 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.444689 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.444702 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.444711 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:47Z","lastTransitionTime":"2026-03-18T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.463026 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.491554 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.509769 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.524368 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.542400 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.551145 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.551183 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.551194 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.551209 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.551222 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:47Z","lastTransitionTime":"2026-03-18T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.566456 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.580347 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.594433 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.607734 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.657452 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.657521 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.657540 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.657565 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.657582 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:47Z","lastTransitionTime":"2026-03-18T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.760050 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.760106 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.760116 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.760128 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.760137 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:47Z","lastTransitionTime":"2026-03-18T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.844151 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.844184 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.844193 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:47 crc kubenswrapper[4983]: E0318 00:09:47.844334 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:47 crc kubenswrapper[4983]: E0318 00:09:47.844473 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:47 crc kubenswrapper[4983]: E0318 00:09:47.844578 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.863322 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.863382 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.863400 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.863423 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.863442 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:47Z","lastTransitionTime":"2026-03-18T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.966541 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.966601 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.966619 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.966649 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:47 crc kubenswrapper[4983]: I0318 00:09:47.966666 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:47Z","lastTransitionTime":"2026-03-18T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.069591 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.069658 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.069683 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.069715 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.069739 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:48Z","lastTransitionTime":"2026-03-18T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.173033 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.173137 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.173157 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.173231 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.173253 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:48Z","lastTransitionTime":"2026-03-18T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.275908 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.276013 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.276037 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.276067 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.276092 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:48Z","lastTransitionTime":"2026-03-18T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.336622 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerStarted","Data":"9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311"} Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.340286 4983 generic.go:334] "Generic (PLEG): container finished" podID="72a1209f-2f3a-48e3-ade2-7aa5824f45eb" containerID="9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b" exitCode=0 Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.340334 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" event={"ID":"72a1209f-2f3a-48e3-ade2-7aa5824f45eb","Type":"ContainerDied","Data":"9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b"} Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.367787 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:48Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.379624 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.379664 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.379677 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.379696 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.379712 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:48Z","lastTransitionTime":"2026-03-18T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.393469 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:48Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.422901 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:48Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.439873 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:48Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.456631 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:48Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.471277 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:48Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.481483 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.481517 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.481527 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.481541 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.481549 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:48Z","lastTransitionTime":"2026-03-18T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.488672 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:48Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.504804 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:48Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.522086 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:48Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.535132 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:48Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.568857 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:48Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.583974 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.584038 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.584055 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.584078 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.584096 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:48Z","lastTransitionTime":"2026-03-18T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.584502 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:48Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.599685 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:48Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.687148 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.687438 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.687548 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.687680 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.687788 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:48Z","lastTransitionTime":"2026-03-18T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.790843 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.790911 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.790963 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.790994 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.791018 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:48Z","lastTransitionTime":"2026-03-18T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.894470 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.894533 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.894551 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.894577 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.894594 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:48Z","lastTransitionTime":"2026-03-18T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.998730 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.999220 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.999272 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.999296 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:48 crc kubenswrapper[4983]: I0318 00:09:48.999313 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:48Z","lastTransitionTime":"2026-03-18T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.101977 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.102021 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.102032 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.102076 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.102090 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:49Z","lastTransitionTime":"2026-03-18T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.205027 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.205090 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.205110 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.205137 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.205167 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:49Z","lastTransitionTime":"2026-03-18T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.309043 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.309103 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.309124 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.309148 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.309165 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:49Z","lastTransitionTime":"2026-03-18T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.348546 4983 generic.go:334] "Generic (PLEG): container finished" podID="72a1209f-2f3a-48e3-ade2-7aa5824f45eb" containerID="4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f" exitCode=0 Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.348633 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" event={"ID":"72a1209f-2f3a-48e3-ade2-7aa5824f45eb","Type":"ContainerDied","Data":"4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f"} Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.398378 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.417766 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.417845 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.417869 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.417898 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.417919 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:49Z","lastTransitionTime":"2026-03-18T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.423783 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.445662 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.461770 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.481920 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.514804 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.521233 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.521280 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.521294 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.521311 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.521323 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:49Z","lastTransitionTime":"2026-03-18T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.532664 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.549748 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.560556 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.575859 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.588916 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.606238 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.619269 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.624117 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.624174 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.624183 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.624199 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.624210 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:49Z","lastTransitionTime":"2026-03-18T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.726024 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.726060 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.726070 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.726083 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.726093 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:49Z","lastTransitionTime":"2026-03-18T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.748300 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-q9xgh"] Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.748713 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-q9xgh" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.751906 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.752388 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.752454 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.752500 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.769044 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aaf037a1-6720-45d2-ada2-bb328bd16e04-host\") pod \"node-ca-q9xgh\" (UID: \"aaf037a1-6720-45d2-ada2-bb328bd16e04\") " pod="openshift-image-registry/node-ca-q9xgh" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.769084 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/aaf037a1-6720-45d2-ada2-bb328bd16e04-serviceca\") pod \"node-ca-q9xgh\" (UID: \"aaf037a1-6720-45d2-ada2-bb328bd16e04\") " pod="openshift-image-registry/node-ca-q9xgh" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.769209 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnmsj\" (UniqueName: \"kubernetes.io/projected/aaf037a1-6720-45d2-ada2-bb328bd16e04-kube-api-access-wnmsj\") pod \"node-ca-q9xgh\" (UID: \"aaf037a1-6720-45d2-ada2-bb328bd16e04\") " pod="openshift-image-registry/node-ca-q9xgh" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.780707 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.794734 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.807282 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.818642 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.829072 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.829117 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.829135 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.829160 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.829177 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:49Z","lastTransitionTime":"2026-03-18T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.829370 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.841704 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.843945 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:49 crc kubenswrapper[4983]: E0318 00:09:49.844066 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.844274 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.844338 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:49 crc kubenswrapper[4983]: E0318 00:09:49.844472 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:49 crc kubenswrapper[4983]: E0318 00:09:49.844747 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.845120 4983 scope.go:117] "RemoveContainer" containerID="b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.860879 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.870001 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aaf037a1-6720-45d2-ada2-bb328bd16e04-host\") pod \"node-ca-q9xgh\" (UID: \"aaf037a1-6720-45d2-ada2-bb328bd16e04\") " pod="openshift-image-registry/node-ca-q9xgh" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.870035 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/aaf037a1-6720-45d2-ada2-bb328bd16e04-serviceca\") pod \"node-ca-q9xgh\" (UID: \"aaf037a1-6720-45d2-ada2-bb328bd16e04\") " pod="openshift-image-registry/node-ca-q9xgh" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.870077 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnmsj\" (UniqueName: \"kubernetes.io/projected/aaf037a1-6720-45d2-ada2-bb328bd16e04-kube-api-access-wnmsj\") pod \"node-ca-q9xgh\" (UID: \"aaf037a1-6720-45d2-ada2-bb328bd16e04\") " pod="openshift-image-registry/node-ca-q9xgh" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.870169 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aaf037a1-6720-45d2-ada2-bb328bd16e04-host\") pod \"node-ca-q9xgh\" (UID: \"aaf037a1-6720-45d2-ada2-bb328bd16e04\") " pod="openshift-image-registry/node-ca-q9xgh" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.872723 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.878159 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/aaf037a1-6720-45d2-ada2-bb328bd16e04-serviceca\") pod \"node-ca-q9xgh\" (UID: \"aaf037a1-6720-45d2-ada2-bb328bd16e04\") " pod="openshift-image-registry/node-ca-q9xgh" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.884696 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnmsj\" (UniqueName: \"kubernetes.io/projected/aaf037a1-6720-45d2-ada2-bb328bd16e04-kube-api-access-wnmsj\") pod \"node-ca-q9xgh\" (UID: \"aaf037a1-6720-45d2-ada2-bb328bd16e04\") " pod="openshift-image-registry/node-ca-q9xgh" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.884812 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.899647 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.920730 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.932558 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.932660 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.932682 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.932716 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.932738 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:49Z","lastTransitionTime":"2026-03-18T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.938279 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.951567 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:49 crc kubenswrapper[4983]: I0318 00:09:49.965791 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.036201 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.036236 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.036246 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.036261 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.036274 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.066443 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-q9xgh" Mar 18 00:09:50 crc kubenswrapper[4983]: W0318 00:09:50.077086 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaaf037a1_6720_45d2_ada2_bb328bd16e04.slice/crio-d8a93941d77c2d84d7d8cbc2ff6c42dcfd86afc229cbedfd804f4a4b7f61f5d6 WatchSource:0}: Error finding container d8a93941d77c2d84d7d8cbc2ff6c42dcfd86afc229cbedfd804f4a4b7f61f5d6: Status 404 returned error can't find the container with id d8a93941d77c2d84d7d8cbc2ff6c42dcfd86afc229cbedfd804f4a4b7f61f5d6 Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.139354 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.140122 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.140138 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.140154 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.140163 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.159074 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.159144 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.159161 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.159212 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.159225 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: E0318 00:09:50.185489 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.193314 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.193355 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.193365 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.193383 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.193392 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: E0318 00:09:50.212122 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.217320 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.217394 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.217421 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.217453 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.217479 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: E0318 00:09:50.236073 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.243451 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.243591 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.243633 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.243650 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.243661 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: E0318 00:09:50.256768 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.260172 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.260287 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.260298 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.260315 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.260334 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: E0318 00:09:50.271353 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: E0318 00:09:50.271499 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.276162 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.276230 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.276250 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.276277 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.276297 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.357385 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.361064 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f"} Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.361493 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.367672 4983 generic.go:334] "Generic (PLEG): container finished" podID="72a1209f-2f3a-48e3-ade2-7aa5824f45eb" containerID="7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7" exitCode=0 Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.367810 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" event={"ID":"72a1209f-2f3a-48e3-ade2-7aa5824f45eb","Type":"ContainerDied","Data":"7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7"} Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.369712 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-q9xgh" event={"ID":"aaf037a1-6720-45d2-ada2-bb328bd16e04","Type":"ContainerStarted","Data":"d8a93941d77c2d84d7d8cbc2ff6c42dcfd86afc229cbedfd804f4a4b7f61f5d6"} Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.380638 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.380680 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.380698 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.380719 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.380736 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.386683 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.388463 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerStarted","Data":"e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e"} Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.389450 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.390794 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.390840 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.419602 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.431846 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.445223 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.464089 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.467306 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.485906 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.486817 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.486876 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.486894 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.486918 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.486987 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.501133 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.516070 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.530486 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.549974 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.569438 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.582430 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.589542 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.589585 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.589604 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.589627 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.589644 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.600303 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.612046 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.634782 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.648237 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.665204 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.679481 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.692160 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.692207 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.692222 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.692245 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.692262 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.705621 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.720789 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.735276 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.757495 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.773184 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.787913 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.794469 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.794525 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.794543 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.794567 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.794584 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.804877 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.822253 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.839252 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.863736 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.883894 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.896643 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.896697 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.896719 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.896747 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.896772 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.999692 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.999744 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:50 crc kubenswrapper[4983]: I0318 00:09:50.999758 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:50.999774 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:50.999789 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:50Z","lastTransitionTime":"2026-03-18T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.103097 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.103156 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.103174 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.103197 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.103215 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:51Z","lastTransitionTime":"2026-03-18T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.205829 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.205890 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.205913 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.205975 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.205999 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:51Z","lastTransitionTime":"2026-03-18T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.308016 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.308042 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.308050 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.308062 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.308070 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:51Z","lastTransitionTime":"2026-03-18T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.394256 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-q9xgh" event={"ID":"aaf037a1-6720-45d2-ada2-bb328bd16e04","Type":"ContainerStarted","Data":"7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7"} Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.400316 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" event={"ID":"72a1209f-2f3a-48e3-ade2-7aa5824f45eb","Type":"ContainerStarted","Data":"25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386"} Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.410835 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.410869 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.410882 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.410899 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.410910 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:51Z","lastTransitionTime":"2026-03-18T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.415899 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.437554 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.455267 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.481548 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.501462 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.513916 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.513972 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.513984 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.513998 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.514009 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:51Z","lastTransitionTime":"2026-03-18T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.532450 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.553455 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.575068 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.597675 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.614921 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.616010 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.616035 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.616044 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.616057 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.616067 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:51Z","lastTransitionTime":"2026-03-18T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.629392 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.643912 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.661188 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.677409 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.693436 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.706913 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.721649 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.721695 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.721707 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.721724 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.721736 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:51Z","lastTransitionTime":"2026-03-18T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.725098 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.746987 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.767978 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.784348 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.805990 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.820973 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.823798 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.823841 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.823858 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.823883 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.823900 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:51Z","lastTransitionTime":"2026-03-18T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.844204 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.844275 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.844212 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:51 crc kubenswrapper[4983]: E0318 00:09:51.844373 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:51 crc kubenswrapper[4983]: E0318 00:09:51.844482 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:51 crc kubenswrapper[4983]: E0318 00:09:51.844593 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.851720 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.867956 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.882593 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.898308 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.913803 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.926436 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.926537 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.926553 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.926570 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.926902 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:51Z","lastTransitionTime":"2026-03-18T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:51 crc kubenswrapper[4983]: I0318 00:09:51.945648 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.030453 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.030500 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.030514 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.030532 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.030545 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:52Z","lastTransitionTime":"2026-03-18T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.132693 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.132726 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.132734 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.132748 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.132757 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:52Z","lastTransitionTime":"2026-03-18T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.234988 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.235019 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.235029 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.235041 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.235050 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:52Z","lastTransitionTime":"2026-03-18T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.337757 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.337807 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.337818 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.337875 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.337887 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:52Z","lastTransitionTime":"2026-03-18T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.440471 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.440513 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.440526 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.440541 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.440552 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:52Z","lastTransitionTime":"2026-03-18T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.543637 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.543702 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.543727 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.543755 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.543773 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:52Z","lastTransitionTime":"2026-03-18T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.646233 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.646309 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.646328 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.646355 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.646375 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:52Z","lastTransitionTime":"2026-03-18T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.749457 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.749506 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.749519 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.749538 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.749551 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:52Z","lastTransitionTime":"2026-03-18T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.851686 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.851736 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.851758 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.851790 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.851815 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:52Z","lastTransitionTime":"2026-03-18T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.868891 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:52Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.887885 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:52Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.917724 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:52Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.935866 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:52Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.956043 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.956091 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.956107 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.956131 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.956149 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:52Z","lastTransitionTime":"2026-03-18T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.961024 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:52Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:52 crc kubenswrapper[4983]: I0318 00:09:52.983392 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:52Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.008653 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.027488 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.049892 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.058570 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.058629 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.058644 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.058662 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.058675 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:53Z","lastTransitionTime":"2026-03-18T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.066258 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.088904 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.104035 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.122706 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.139836 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.160713 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.160748 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.160758 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.160771 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.160781 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:53Z","lastTransitionTime":"2026-03-18T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.263414 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.263482 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.263500 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.263525 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.263542 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:53Z","lastTransitionTime":"2026-03-18T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.367247 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.367329 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.367343 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.367361 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.367670 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:53Z","lastTransitionTime":"2026-03-18T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.408644 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/0.log" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.412396 4983 generic.go:334] "Generic (PLEG): container finished" podID="938043fe-450c-4477-8f55-312a843b0706" containerID="e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e" exitCode=1 Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.412452 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e"} Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.413514 4983 scope.go:117] "RemoveContainer" containerID="e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.430801 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.454203 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.471391 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.471458 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.471480 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.471510 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.471531 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:53Z","lastTransitionTime":"2026-03-18T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.473362 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.502981 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.517660 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.533390 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.555309 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:09:52Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:52.776913 6792 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:09:52.777072 6792 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0318 00:09:52.777115 6792 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0318 00:09:52.777137 6792 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0318 00:09:52.777168 6792 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0318 00:09:52.777204 6792 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0318 00:09:52.777228 6792 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0318 00:09:52.777260 6792 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0318 00:09:52.777273 6792 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0318 00:09:52.777283 6792 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0318 00:09:52.777311 6792 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.570336 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.574329 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.574380 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.574394 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.574415 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.574434 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:53Z","lastTransitionTime":"2026-03-18T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.586746 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.600533 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.613131 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.625513 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.642734 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.657542 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.676873 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.676904 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.676912 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.676939 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.676952 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:53Z","lastTransitionTime":"2026-03-18T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.779661 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.779712 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.779725 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.779739 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.779748 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:53Z","lastTransitionTime":"2026-03-18T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.848827 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.848883 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:53 crc kubenswrapper[4983]: E0318 00:09:53.848952 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:53 crc kubenswrapper[4983]: E0318 00:09:53.849034 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.849109 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:53 crc kubenswrapper[4983]: E0318 00:09:53.849169 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.881990 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.882022 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.882032 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.882045 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.882054 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:53Z","lastTransitionTime":"2026-03-18T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.984463 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.984503 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.984515 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.984532 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:53 crc kubenswrapper[4983]: I0318 00:09:53.984543 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:53Z","lastTransitionTime":"2026-03-18T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.087172 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.087208 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.087219 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.087235 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.087246 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:54Z","lastTransitionTime":"2026-03-18T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.190301 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.190346 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.190362 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.190386 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.190402 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:54Z","lastTransitionTime":"2026-03-18T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.292785 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.292848 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.292871 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.292897 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.292918 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:54Z","lastTransitionTime":"2026-03-18T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.396089 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.396170 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.396189 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.396214 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.396231 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:54Z","lastTransitionTime":"2026-03-18T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.418007 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/0.log" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.421125 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerStarted","Data":"c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f"} Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.421846 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.441769 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.457984 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.483325 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:09:52Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:52.776913 6792 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:09:52.777072 6792 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0318 00:09:52.777115 6792 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0318 00:09:52.777137 6792 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0318 00:09:52.777168 6792 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0318 00:09:52.777204 6792 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0318 00:09:52.777228 6792 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0318 00:09:52.777260 6792 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0318 00:09:52.777273 6792 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0318 00:09:52.777283 6792 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0318 00:09:52.777311 6792 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.499381 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.499569 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.499686 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.499807 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.499921 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:54Z","lastTransitionTime":"2026-03-18T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.505161 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.524544 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.542001 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.558763 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.576723 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.600434 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.604463 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.604531 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.604550 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.604575 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.604590 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:54Z","lastTransitionTime":"2026-03-18T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.622892 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.638322 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.657341 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.671781 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.703975 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.706489 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.706546 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.706560 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.706581 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.706595 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:54Z","lastTransitionTime":"2026-03-18T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.809381 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.809445 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.809463 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.809489 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.809506 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:54Z","lastTransitionTime":"2026-03-18T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.912524 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.912590 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.912608 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.912632 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:54 crc kubenswrapper[4983]: I0318 00:09:54.912652 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:54Z","lastTransitionTime":"2026-03-18T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.015663 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.015725 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.015745 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.015770 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.015786 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:55Z","lastTransitionTime":"2026-03-18T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.119137 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.119212 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.119231 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.119260 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.119277 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:55Z","lastTransitionTime":"2026-03-18T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.256698 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.256764 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.256780 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.256804 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.256822 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:55Z","lastTransitionTime":"2026-03-18T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.360308 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.360369 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.360385 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.360410 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.360428 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:55Z","lastTransitionTime":"2026-03-18T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.427525 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/1.log" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.428506 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/0.log" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.432456 4983 generic.go:334] "Generic (PLEG): container finished" podID="938043fe-450c-4477-8f55-312a843b0706" containerID="c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f" exitCode=1 Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.432527 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f"} Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.432586 4983 scope.go:117] "RemoveContainer" containerID="e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.433884 4983 scope.go:117] "RemoveContainer" containerID="c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f" Mar 18 00:09:55 crc kubenswrapper[4983]: E0318 00:09:55.435165 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.459694 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.463332 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.463656 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.463667 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.463685 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.463696 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:55Z","lastTransitionTime":"2026-03-18T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.476060 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.492960 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.523153 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.544853 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.567827 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.567874 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.567892 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.567859 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.567915 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.568137 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:55Z","lastTransitionTime":"2026-03-18T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.587489 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.603346 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.633536 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.649255 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.670044 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.671530 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.671634 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.671656 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.671680 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.671698 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:55Z","lastTransitionTime":"2026-03-18T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.691082 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.711154 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.744538 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:09:52Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:52.776913 6792 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:09:52.777072 6792 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0318 00:09:52.777115 6792 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0318 00:09:52.777137 6792 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0318 00:09:52.777168 6792 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0318 00:09:52.777204 6792 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0318 00:09:52.777228 6792 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0318 00:09:52.777260 6792 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0318 00:09:52.777273 6792 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0318 00:09:52.777283 6792 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0318 00:09:52.777311 6792 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:09:54Z\\\",\\\"message\\\":\\\"go:160\\\\nI0318 00:09:54.389534 6980 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389565 6980 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389673 6980 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389893 6980 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390106 6980 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390199 6980 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:09:54.390236 6980 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0318 00:09:54.390377 6980 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.774247 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.774292 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.774308 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.774331 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.774348 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:55Z","lastTransitionTime":"2026-03-18T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.807729 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2"] Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.808419 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:55 crc kubenswrapper[4983]: W0318 00:09:55.811651 4983 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd": failed to list *v1.Secret: secrets "ovn-kubernetes-control-plane-dockercfg-gs7dd" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Mar 18 00:09:55 crc kubenswrapper[4983]: E0318 00:09:55.811717 4983 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-control-plane-dockercfg-gs7dd\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-control-plane-dockercfg-gs7dd\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 00:09:55 crc kubenswrapper[4983]: W0318 00:09:55.811785 4983 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert": failed to list *v1.Secret: secrets "ovn-control-plane-metrics-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Mar 18 00:09:55 crc kubenswrapper[4983]: E0318 00:09:55.811808 4983 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-control-plane-metrics-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-control-plane-metrics-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.839459 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.843539 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.843551 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:55 crc kubenswrapper[4983]: E0318 00:09:55.843678 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.843793 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:55 crc kubenswrapper[4983]: E0318 00:09:55.844004 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.844333 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh2c8\" (UniqueName: \"kubernetes.io/projected/2fa65674-b13b-4d23-8fe9-9b677435bcd2-kube-api-access-xh2c8\") pod \"ovnkube-control-plane-749d76644c-67wt2\" (UID: \"2fa65674-b13b-4d23-8fe9-9b677435bcd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:55 crc kubenswrapper[4983]: E0318 00:09:55.844326 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.844417 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2fa65674-b13b-4d23-8fe9-9b677435bcd2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-67wt2\" (UID: \"2fa65674-b13b-4d23-8fe9-9b677435bcd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.844453 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2fa65674-b13b-4d23-8fe9-9b677435bcd2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-67wt2\" (UID: \"2fa65674-b13b-4d23-8fe9-9b677435bcd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.844490 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2fa65674-b13b-4d23-8fe9-9b677435bcd2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-67wt2\" (UID: \"2fa65674-b13b-4d23-8fe9-9b677435bcd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.857606 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.877701 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.878111 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.878158 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.878181 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.878211 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.878234 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:55Z","lastTransitionTime":"2026-03-18T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.899640 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.909720 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.919622 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.933246 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.944968 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh2c8\" (UniqueName: \"kubernetes.io/projected/2fa65674-b13b-4d23-8fe9-9b677435bcd2-kube-api-access-xh2c8\") pod \"ovnkube-control-plane-749d76644c-67wt2\" (UID: \"2fa65674-b13b-4d23-8fe9-9b677435bcd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.945036 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2fa65674-b13b-4d23-8fe9-9b677435bcd2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-67wt2\" (UID: \"2fa65674-b13b-4d23-8fe9-9b677435bcd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.945064 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2fa65674-b13b-4d23-8fe9-9b677435bcd2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-67wt2\" (UID: \"2fa65674-b13b-4d23-8fe9-9b677435bcd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.945090 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2fa65674-b13b-4d23-8fe9-9b677435bcd2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-67wt2\" (UID: \"2fa65674-b13b-4d23-8fe9-9b677435bcd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.945821 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2fa65674-b13b-4d23-8fe9-9b677435bcd2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-67wt2\" (UID: \"2fa65674-b13b-4d23-8fe9-9b677435bcd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.945982 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2fa65674-b13b-4d23-8fe9-9b677435bcd2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-67wt2\" (UID: \"2fa65674-b13b-4d23-8fe9-9b677435bcd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.952775 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4425548c3f9855ccb95ffaed668cdff6ac5b2a540454bacb85e540def17734e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:09:52Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:52.776913 6792 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:09:52.777072 6792 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0318 00:09:52.777115 6792 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0318 00:09:52.777137 6792 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0318 00:09:52.777168 6792 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0318 00:09:52.777204 6792 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0318 00:09:52.777228 6792 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0318 00:09:52.777260 6792 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0318 00:09:52.777273 6792 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0318 00:09:52.777283 6792 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0318 00:09:52.777311 6792 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:09:54Z\\\",\\\"message\\\":\\\"go:160\\\\nI0318 00:09:54.389534 6980 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389565 6980 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389673 6980 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389893 6980 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390106 6980 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390199 6980 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:09:54.390236 6980 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0318 00:09:54.390377 6980 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.966633 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh2c8\" (UniqueName: \"kubernetes.io/projected/2fa65674-b13b-4d23-8fe9-9b677435bcd2-kube-api-access-xh2c8\") pod \"ovnkube-control-plane-749d76644c-67wt2\" (UID: \"2fa65674-b13b-4d23-8fe9-9b677435bcd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.968730 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.980518 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.980547 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.980559 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.980575 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.980587 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:55Z","lastTransitionTime":"2026-03-18T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.981273 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:55 crc kubenswrapper[4983]: I0318 00:09:55.992185 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:55Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.008372 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.023502 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.036288 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.047160 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.083272 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.083321 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.083338 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.083362 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.083380 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:56Z","lastTransitionTime":"2026-03-18T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.187086 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.187141 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.187158 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.187182 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.187199 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:56Z","lastTransitionTime":"2026-03-18T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.290104 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.290162 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.290178 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.290202 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.290219 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:56Z","lastTransitionTime":"2026-03-18T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.393915 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.394242 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.394369 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.394544 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.394669 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:56Z","lastTransitionTime":"2026-03-18T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.439339 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/1.log" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.444633 4983 scope.go:117] "RemoveContainer" containerID="c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f" Mar 18 00:09:56 crc kubenswrapper[4983]: E0318 00:09:56.444947 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.465527 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.484891 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.497976 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.498037 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.498054 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.498078 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.498095 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:56Z","lastTransitionTime":"2026-03-18T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.519551 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:09:54Z\\\",\\\"message\\\":\\\"go:160\\\\nI0318 00:09:54.389534 6980 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389565 6980 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389673 6980 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389893 6980 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390106 6980 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390199 6980 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:09:54.390236 6980 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0318 00:09:54.390377 6980 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.539049 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.562897 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.564837 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-vjcg5"] Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.565756 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:09:56 crc kubenswrapper[4983]: E0318 00:09:56.566028 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.583329 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.601239 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.601295 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.601358 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.601390 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.601408 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:56Z","lastTransitionTime":"2026-03-18T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.601980 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.619650 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.636054 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.651244 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.653804 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.653869 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk9q9\" (UniqueName: \"kubernetes.io/projected/fe22dcea-cd03-43e3-80da-61a5f1e063ad-kube-api-access-mk9q9\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.664188 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.677589 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.690689 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.704758 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.704809 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.704820 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.704836 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.704847 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:56Z","lastTransitionTime":"2026-03-18T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.708814 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.722720 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.737119 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.751760 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.754725 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.755029 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk9q9\" (UniqueName: \"kubernetes.io/projected/fe22dcea-cd03-43e3-80da-61a5f1e063ad-kube-api-access-mk9q9\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:09:56 crc kubenswrapper[4983]: E0318 00:09:56.754917 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:09:56 crc kubenswrapper[4983]: E0318 00:09:56.755362 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs podName:fe22dcea-cd03-43e3-80da-61a5f1e063ad nodeName:}" failed. No retries permitted until 2026-03-18 00:09:57.255337182 +0000 UTC m=+115.152711802 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs") pod "network-metrics-daemon-vjcg5" (UID: "fe22dcea-cd03-43e3-80da-61a5f1e063ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.772525 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.784590 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk9q9\" (UniqueName: \"kubernetes.io/projected/fe22dcea-cd03-43e3-80da-61a5f1e063ad-kube-api-access-mk9q9\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.789385 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.807193 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.807238 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.807251 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.807335 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.807364 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:56Z","lastTransitionTime":"2026-03-18T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.813836 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:09:54Z\\\",\\\"message\\\":\\\"go:160\\\\nI0318 00:09:54.389534 6980 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389565 6980 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389673 6980 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389893 6980 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390106 6980 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390199 6980 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:09:54.390236 6980 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0318 00:09:54.390377 6980 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.833324 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.852148 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.868990 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.890765 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.911615 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.911657 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.911670 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.911688 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.911700 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:56Z","lastTransitionTime":"2026-03-18T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.912116 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.930955 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: E0318 00:09:56.946230 4983 secret.go:188] Couldn't get secret openshift-ovn-kubernetes/ovn-control-plane-metrics-cert: failed to sync secret cache: timed out waiting for the condition Mar 18 00:09:56 crc kubenswrapper[4983]: E0318 00:09:56.946630 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2fa65674-b13b-4d23-8fe9-9b677435bcd2-ovn-control-plane-metrics-cert podName:2fa65674-b13b-4d23-8fe9-9b677435bcd2 nodeName:}" failed. No retries permitted until 2026-03-18 00:09:57.446600146 +0000 UTC m=+115.343974786 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovn-control-plane-metrics-cert" (UniqueName: "kubernetes.io/secret/2fa65674-b13b-4d23-8fe9-9b677435bcd2-ovn-control-plane-metrics-cert") pod "ovnkube-control-plane-749d76644c-67wt2" (UID: "2fa65674-b13b-4d23-8fe9-9b677435bcd2") : failed to sync secret cache: timed out waiting for the condition Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.951284 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.967892 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:56 crc kubenswrapper[4983]: I0318 00:09:56.999050 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:56Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.012045 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.013700 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.013745 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.013757 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.013774 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.013786 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:57Z","lastTransitionTime":"2026-03-18T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.033567 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.093137 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.115822 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.115872 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.115886 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.115904 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.115917 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:57Z","lastTransitionTime":"2026-03-18T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.218700 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.218769 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.218788 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.218811 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.218829 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:57Z","lastTransitionTime":"2026-03-18T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.259999 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:09:57 crc kubenswrapper[4983]: E0318 00:09:57.260271 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:09:57 crc kubenswrapper[4983]: E0318 00:09:57.260378 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs podName:fe22dcea-cd03-43e3-80da-61a5f1e063ad nodeName:}" failed. No retries permitted until 2026-03-18 00:09:58.260351401 +0000 UTC m=+116.157726051 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs") pod "network-metrics-daemon-vjcg5" (UID: "fe22dcea-cd03-43e3-80da-61a5f1e063ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.270588 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.321798 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.321848 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.321865 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.321887 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.321904 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:57Z","lastTransitionTime":"2026-03-18T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.424710 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.424794 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.424823 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.424857 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.424880 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:57Z","lastTransitionTime":"2026-03-18T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.461595 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2fa65674-b13b-4d23-8fe9-9b677435bcd2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-67wt2\" (UID: \"2fa65674-b13b-4d23-8fe9-9b677435bcd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.466600 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2fa65674-b13b-4d23-8fe9-9b677435bcd2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-67wt2\" (UID: \"2fa65674-b13b-4d23-8fe9-9b677435bcd2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.528208 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.528257 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.528273 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.528297 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.528314 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:57Z","lastTransitionTime":"2026-03-18T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.635384 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.638323 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.638363 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.638378 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.638400 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.638418 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:57Z","lastTransitionTime":"2026-03-18T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:57 crc kubenswrapper[4983]: W0318 00:09:57.656569 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fa65674_b13b_4d23_8fe9_9b677435bcd2.slice/crio-644f25f3cfa97cef4bfadd466388764d018d2f3e6517b2424700ccb969fac496 WatchSource:0}: Error finding container 644f25f3cfa97cef4bfadd466388764d018d2f3e6517b2424700ccb969fac496: Status 404 returned error can't find the container with id 644f25f3cfa97cef4bfadd466388764d018d2f3e6517b2424700ccb969fac496 Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.741888 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.741974 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.741994 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.742018 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.742036 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:57Z","lastTransitionTime":"2026-03-18T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.843558 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:57 crc kubenswrapper[4983]: E0318 00:09:57.843726 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.843572 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.843572 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:09:57 crc kubenswrapper[4983]: E0318 00:09:57.843848 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:57 crc kubenswrapper[4983]: E0318 00:09:57.843986 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.844690 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:57 crc kubenswrapper[4983]: E0318 00:09:57.844870 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.848749 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.848788 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.848803 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.848862 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.848882 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:57Z","lastTransitionTime":"2026-03-18T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.952314 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.952676 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.952714 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.952753 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:57 crc kubenswrapper[4983]: I0318 00:09:57.952766 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:57Z","lastTransitionTime":"2026-03-18T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.055343 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.055450 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.055460 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.055477 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.055489 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:58Z","lastTransitionTime":"2026-03-18T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.159666 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.159733 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.159751 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.159779 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.159798 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:58Z","lastTransitionTime":"2026-03-18T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.262274 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.262342 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.262361 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.262385 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.262402 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:58Z","lastTransitionTime":"2026-03-18T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.300896 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:09:58 crc kubenswrapper[4983]: E0318 00:09:58.302077 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:09:58 crc kubenswrapper[4983]: E0318 00:09:58.302167 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs podName:fe22dcea-cd03-43e3-80da-61a5f1e063ad nodeName:}" failed. No retries permitted until 2026-03-18 00:10:00.302150181 +0000 UTC m=+118.199524791 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs") pod "network-metrics-daemon-vjcg5" (UID: "fe22dcea-cd03-43e3-80da-61a5f1e063ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.365149 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.365211 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.365233 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.365266 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.365290 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:58Z","lastTransitionTime":"2026-03-18T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.453956 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" event={"ID":"2fa65674-b13b-4d23-8fe9-9b677435bcd2","Type":"ContainerStarted","Data":"5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8"} Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.454040 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" event={"ID":"2fa65674-b13b-4d23-8fe9-9b677435bcd2","Type":"ContainerStarted","Data":"064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c"} Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.454068 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" event={"ID":"2fa65674-b13b-4d23-8fe9-9b677435bcd2","Type":"ContainerStarted","Data":"644f25f3cfa97cef4bfadd466388764d018d2f3e6517b2424700ccb969fac496"} Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.468026 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.468082 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.468101 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.468123 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.468139 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:58Z","lastTransitionTime":"2026-03-18T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.494465 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.518374 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.539318 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.557724 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.571376 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.571442 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.571465 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.571494 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.571520 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:58Z","lastTransitionTime":"2026-03-18T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.579450 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.599573 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.625199 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.659087 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:09:54Z\\\",\\\"message\\\":\\\"go:160\\\\nI0318 00:09:54.389534 6980 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389565 6980 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389673 6980 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389893 6980 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390106 6980 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390199 6980 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:09:54.390236 6980 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0318 00:09:54.390377 6980 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.675263 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.675345 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.675364 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.675776 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.675834 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:58Z","lastTransitionTime":"2026-03-18T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.680131 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.699967 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.725158 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.747549 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.769114 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.779013 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.779055 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.779071 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.779093 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.779109 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:58Z","lastTransitionTime":"2026-03-18T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.789327 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.809270 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.827183 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:09:58Z is after 2025-08-24T17:21:41Z" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.882537 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.882592 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.882612 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.882636 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.882654 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:58Z","lastTransitionTime":"2026-03-18T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.985578 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.985659 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.985678 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.985701 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:58 crc kubenswrapper[4983]: I0318 00:09:58.985718 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:58Z","lastTransitionTime":"2026-03-18T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.088550 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.088617 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.088641 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.088669 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.088691 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:59Z","lastTransitionTime":"2026-03-18T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.193229 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.193282 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.193307 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.193327 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.193342 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:59Z","lastTransitionTime":"2026-03-18T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.295654 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.295685 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.295694 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.295706 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.295714 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:59Z","lastTransitionTime":"2026-03-18T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.398349 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.398423 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.398448 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.398479 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.398502 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:59Z","lastTransitionTime":"2026-03-18T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.500914 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.501013 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.501030 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.501056 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.501074 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:59Z","lastTransitionTime":"2026-03-18T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.603528 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.603599 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.603622 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.603652 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.603677 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:59Z","lastTransitionTime":"2026-03-18T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.706707 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.706769 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.706786 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.706810 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.706828 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:59Z","lastTransitionTime":"2026-03-18T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.717566 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.717681 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.717725 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.717756 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:10:31.717719159 +0000 UTC m=+149.615093809 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.717811 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.717893 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.717911 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.717964 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.717991 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.718027 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.717897 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.718034 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:10:31.717999567 +0000 UTC m=+149.615374217 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.718100 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 00:10:31.71808594 +0000 UTC m=+149.615460590 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.718125 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:10:31.718111851 +0000 UTC m=+149.615486491 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.718120 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.718167 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.718185 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.718253 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 00:10:31.718233884 +0000 UTC m=+149.615608524 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.809756 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.809815 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.809837 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.809868 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.809890 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:59Z","lastTransitionTime":"2026-03-18T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.844135 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.844196 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.844228 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.844197 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.844322 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.844537 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.844634 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:09:59 crc kubenswrapper[4983]: E0318 00:09:59.844747 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.913384 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.913463 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.913480 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.913509 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:09:59 crc kubenswrapper[4983]: I0318 00:09:59.913527 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:09:59Z","lastTransitionTime":"2026-03-18T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.016469 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.016529 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.016546 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.016569 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.016591 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.119603 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.119685 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.119712 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.119744 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.119767 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.222717 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.222781 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.222799 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.222824 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.222847 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.328082 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:00 crc kubenswrapper[4983]: E0318 00:10:00.328300 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:10:00 crc kubenswrapper[4983]: E0318 00:10:00.328408 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs podName:fe22dcea-cd03-43e3-80da-61a5f1e063ad nodeName:}" failed. No retries permitted until 2026-03-18 00:10:04.328375498 +0000 UTC m=+122.225750148 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs") pod "network-metrics-daemon-vjcg5" (UID: "fe22dcea-cd03-43e3-80da-61a5f1e063ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.328738 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.328793 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.328815 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.328844 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.328865 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.330290 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.330339 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.330361 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.330386 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.330406 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: E0318 00:10:00.344618 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:00Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.348856 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.348898 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.348907 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.348921 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.348944 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: E0318 00:10:00.367394 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:00Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.371580 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.371640 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.371674 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.371698 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.371715 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: E0318 00:10:00.390486 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:00Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.394558 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.394638 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.394669 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.394702 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.394726 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: E0318 00:10:00.407096 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:00Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.410886 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.411028 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.411040 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.411052 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.411060 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: E0318 00:10:00.422273 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:00Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:00 crc kubenswrapper[4983]: E0318 00:10:00.422579 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.431163 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.431261 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.431308 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.431322 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.431331 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.533530 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.533611 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.533631 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.533654 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.533671 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.636598 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.636670 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.636693 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.636722 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.636747 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.740130 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.740207 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.740229 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.740257 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.740281 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.827836 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.843724 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.843783 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.843805 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.843834 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.843856 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.851568 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:00Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.874761 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:00Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.894009 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:00Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.925250 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:09:54Z\\\",\\\"message\\\":\\\"go:160\\\\nI0318 00:09:54.389534 6980 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389565 6980 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389673 6980 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389893 6980 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390106 6980 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390199 6980 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:09:54.390236 6980 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0318 00:09:54.390377 6980 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:00Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.942721 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:00Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.947772 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.947833 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.947852 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.947877 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.947894 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:00Z","lastTransitionTime":"2026-03-18T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.960397 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:00Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:00 crc kubenswrapper[4983]: I0318 00:10:00.982880 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:00Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.005524 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:01Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.026160 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:01Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.044889 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:01Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.050666 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.050731 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.050750 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.050775 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.050792 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:01Z","lastTransitionTime":"2026-03-18T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.063717 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:01Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.081107 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:01Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.114713 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:01Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.132209 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:01Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.153451 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:01Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.154328 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.154379 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.154415 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.154449 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.154471 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:01Z","lastTransitionTime":"2026-03-18T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.169055 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:01Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.258819 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.258891 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.258915 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.258978 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.259002 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:01Z","lastTransitionTime":"2026-03-18T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.361963 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.362035 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.362055 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.362077 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.362093 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:01Z","lastTransitionTime":"2026-03-18T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.466591 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.466648 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.466665 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.466688 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.466705 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:01Z","lastTransitionTime":"2026-03-18T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.570816 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.570883 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.570896 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.570919 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.570955 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:01Z","lastTransitionTime":"2026-03-18T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.674433 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.674508 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.674535 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.674565 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.674586 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:01Z","lastTransitionTime":"2026-03-18T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.778708 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.778797 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.778817 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.778839 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.778856 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:01Z","lastTransitionTime":"2026-03-18T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.844138 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:01 crc kubenswrapper[4983]: E0318 00:10:01.844332 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.844853 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:01 crc kubenswrapper[4983]: E0318 00:10:01.845208 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.845296 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.845258 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:01 crc kubenswrapper[4983]: E0318 00:10:01.845499 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:01 crc kubenswrapper[4983]: E0318 00:10:01.845746 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.882003 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.882056 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.882073 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.882096 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.882113 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:01Z","lastTransitionTime":"2026-03-18T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.985974 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.986035 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.986052 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.986080 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:01 crc kubenswrapper[4983]: I0318 00:10:01.986098 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:01Z","lastTransitionTime":"2026-03-18T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.089261 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.089299 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.089310 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.089324 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.089333 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:02Z","lastTransitionTime":"2026-03-18T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.192258 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.192318 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.192334 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.192358 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.192375 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:02Z","lastTransitionTime":"2026-03-18T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.295463 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.295536 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.295562 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.295593 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.295615 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:02Z","lastTransitionTime":"2026-03-18T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.398714 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.398775 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.398794 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.398818 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.398836 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:02Z","lastTransitionTime":"2026-03-18T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.502270 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.502353 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.502376 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.502406 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.502428 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:02Z","lastTransitionTime":"2026-03-18T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.606378 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.606466 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.606483 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.606550 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.606569 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:02Z","lastTransitionTime":"2026-03-18T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.710007 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.710085 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.710108 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.710139 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:02 crc kubenswrapper[4983]: I0318 00:10:02.710162 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:02Z","lastTransitionTime":"2026-03-18T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:03 crc kubenswrapper[4983]: E0318 00:10:02.811203 4983 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:02.877526 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:02.899365 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:02.920028 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:02.939058 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: E0318 00:10:02.949419 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:02.957395 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:02.975554 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:02.998438 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:03.030271 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:03.046339 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:03.066434 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:03.081130 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:03.096955 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:03.116865 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:03.143266 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:03.171436 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:09:54Z\\\",\\\"message\\\":\\\"go:160\\\\nI0318 00:09:54.389534 6980 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389565 6980 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389673 6980 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389893 6980 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390106 6980 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390199 6980 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:09:54.390236 6980 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0318 00:09:54.390377 6980 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:03.187991 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:03.844196 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:03.844278 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:03.844391 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:03 crc kubenswrapper[4983]: E0318 00:10:03.844377 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:03 crc kubenswrapper[4983]: E0318 00:10:03.844575 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:03 crc kubenswrapper[4983]: I0318 00:10:03.844696 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:03 crc kubenswrapper[4983]: E0318 00:10:03.844801 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:03 crc kubenswrapper[4983]: E0318 00:10:03.844898 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:04 crc kubenswrapper[4983]: I0318 00:10:04.378619 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:04 crc kubenswrapper[4983]: E0318 00:10:04.379220 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:10:04 crc kubenswrapper[4983]: E0318 00:10:04.379292 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs podName:fe22dcea-cd03-43e3-80da-61a5f1e063ad nodeName:}" failed. No retries permitted until 2026-03-18 00:10:12.379270936 +0000 UTC m=+130.276645586 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs") pod "network-metrics-daemon-vjcg5" (UID: "fe22dcea-cd03-43e3-80da-61a5f1e063ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:10:05 crc kubenswrapper[4983]: I0318 00:10:05.843859 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:05 crc kubenswrapper[4983]: E0318 00:10:05.845152 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:05 crc kubenswrapper[4983]: I0318 00:10:05.843979 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:05 crc kubenswrapper[4983]: I0318 00:10:05.844062 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:05 crc kubenswrapper[4983]: I0318 00:10:05.843978 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:05 crc kubenswrapper[4983]: E0318 00:10:05.845775 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:05 crc kubenswrapper[4983]: E0318 00:10:05.846182 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:05 crc kubenswrapper[4983]: E0318 00:10:05.846216 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:07 crc kubenswrapper[4983]: I0318 00:10:07.843692 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:07 crc kubenswrapper[4983]: I0318 00:10:07.843719 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:07 crc kubenswrapper[4983]: I0318 00:10:07.843868 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:07 crc kubenswrapper[4983]: E0318 00:10:07.843868 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:07 crc kubenswrapper[4983]: I0318 00:10:07.843968 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:07 crc kubenswrapper[4983]: E0318 00:10:07.844085 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:07 crc kubenswrapper[4983]: E0318 00:10:07.844637 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:07 crc kubenswrapper[4983]: E0318 00:10:07.844798 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:07 crc kubenswrapper[4983]: I0318 00:10:07.845187 4983 scope.go:117] "RemoveContainer" containerID="c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f" Mar 18 00:10:07 crc kubenswrapper[4983]: E0318 00:10:07.951202 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.496120 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/1.log" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.499509 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerStarted","Data":"89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707"} Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.499969 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.523735 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.538325 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.551028 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.562243 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.575621 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.592292 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.609674 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.640466 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:09:54Z\\\",\\\"message\\\":\\\"go:160\\\\nI0318 00:09:54.389534 6980 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389565 6980 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389673 6980 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389893 6980 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390106 6980 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390199 6980 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:09:54.390236 6980 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0318 00:09:54.390377 6980 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.657329 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.672527 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.685591 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.708900 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.726894 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.740620 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.753586 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.769699 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:08Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:08 crc kubenswrapper[4983]: I0318 00:10:08.863379 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.506093 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/2.log" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.506851 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/1.log" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.511424 4983 generic.go:334] "Generic (PLEG): container finished" podID="938043fe-450c-4477-8f55-312a843b0706" containerID="89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707" exitCode=1 Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.511540 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707"} Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.511668 4983 scope.go:117] "RemoveContainer" containerID="c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.512697 4983 scope.go:117] "RemoveContainer" containerID="89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707" Mar 18 00:10:09 crc kubenswrapper[4983]: E0318 00:10:09.513197 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.537434 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.559120 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.579906 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.612255 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c23d33782b58dfacc94ef02b7f72f2c52891e5948ed21304fbb7aa311ad42b3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:09:54Z\\\",\\\"message\\\":\\\"go:160\\\\nI0318 00:09:54.389534 6980 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389565 6980 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389673 6980 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.389893 6980 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390106 6980 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:09:54.390199 6980 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:09:54.390236 6980 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0318 00:09:54.390377 6980 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:08Z\\\",\\\"message\\\":\\\"rts Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 00:10:08.902818 7235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}\\\\nF0318 00:10:08.903148 7235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.632969 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.655047 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.673819 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.694544 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.718391 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.740203 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e68e246-bb9b-4b78-b668-ddcc38a6b26f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:08:35Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 00:08:05.013265 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 00:08:05.015776 1 observer_polling.go:159] Starting file observer\\\\nI0318 00:08:05.046728 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 00:08:05.051577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 00:08:35.404406 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.762696 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.783863 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.801755 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.839515 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.843617 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.843668 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.843710 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.843540 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:09 crc kubenswrapper[4983]: E0318 00:10:09.843838 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:09 crc kubenswrapper[4983]: E0318 00:10:09.844039 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:09 crc kubenswrapper[4983]: E0318 00:10:09.844340 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:09 crc kubenswrapper[4983]: E0318 00:10:09.844911 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.858280 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.882618 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:09 crc kubenswrapper[4983]: I0318 00:10:09.903984 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:09Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.518793 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/2.log" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.526432 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.526683 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.526816 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.526976 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.527139 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:10Z","lastTransitionTime":"2026-03-18T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.530253 4983 scope.go:117] "RemoveContainer" containerID="89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707" Mar 18 00:10:10 crc kubenswrapper[4983]: E0318 00:10:10.530613 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.547513 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: E0318 00:10:10.550095 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.555627 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.555683 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.555700 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.555724 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.555742 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:10Z","lastTransitionTime":"2026-03-18T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.576233 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: E0318 00:10:10.576706 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.581451 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.581643 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.581777 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.582024 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.582195 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:10Z","lastTransitionTime":"2026-03-18T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.601068 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: E0318 00:10:10.602255 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.607588 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.607642 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.607658 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.607684 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.607705 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:10Z","lastTransitionTime":"2026-03-18T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.623882 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e68e246-bb9b-4b78-b668-ddcc38a6b26f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:08:35Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 00:08:05.013265 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 00:08:05.015776 1 observer_polling.go:159] Starting file observer\\\\nI0318 00:08:05.046728 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 00:08:05.051577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 00:08:35.404406 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: E0318 00:10:10.627350 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.632106 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.632163 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.632181 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.632205 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.632222 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:10Z","lastTransitionTime":"2026-03-18T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.646869 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: E0318 00:10:10.655688 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: E0318 00:10:10.655908 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.666982 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.690210 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.714282 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.748141 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.761176 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.778014 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.790882 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.803352 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.818611 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.834416 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.859253 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:08Z\\\",\\\"message\\\":\\\"rts Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 00:10:08.902818 7235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}\\\\nF0318 00:10:08.903148 7235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:10 crc kubenswrapper[4983]: I0318 00:10:10.874452 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:10Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:11 crc kubenswrapper[4983]: I0318 00:10:11.843767 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:11 crc kubenswrapper[4983]: I0318 00:10:11.843843 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:11 crc kubenswrapper[4983]: I0318 00:10:11.843844 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:11 crc kubenswrapper[4983]: I0318 00:10:11.843997 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:11 crc kubenswrapper[4983]: E0318 00:10:11.843981 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:11 crc kubenswrapper[4983]: E0318 00:10:11.844212 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:11 crc kubenswrapper[4983]: E0318 00:10:11.844317 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:11 crc kubenswrapper[4983]: E0318 00:10:11.844393 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:12 crc kubenswrapper[4983]: I0318 00:10:12.473099 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:12 crc kubenswrapper[4983]: E0318 00:10:12.473360 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:10:12 crc kubenswrapper[4983]: E0318 00:10:12.473493 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs podName:fe22dcea-cd03-43e3-80da-61a5f1e063ad nodeName:}" failed. No retries permitted until 2026-03-18 00:10:28.473462323 +0000 UTC m=+146.370836973 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs") pod "network-metrics-daemon-vjcg5" (UID: "fe22dcea-cd03-43e3-80da-61a5f1e063ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:10:12 crc kubenswrapper[4983]: I0318 00:10:12.862044 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:12Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:12 crc kubenswrapper[4983]: I0318 00:10:12.887109 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:12Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:12 crc kubenswrapper[4983]: I0318 00:10:12.923584 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:12Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:12 crc kubenswrapper[4983]: E0318 00:10:12.961513 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:10:12 crc kubenswrapper[4983]: I0318 00:10:12.968168 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:12Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:12 crc kubenswrapper[4983]: I0318 00:10:12.984302 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:12Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:12 crc kubenswrapper[4983]: I0318 00:10:12.997678 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:12Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.014147 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:13Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.037696 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:08Z\\\",\\\"message\\\":\\\"rts Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 00:10:08.902818 7235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}\\\\nF0318 00:10:08.903148 7235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:13Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.051240 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:13Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.067901 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e68e246-bb9b-4b78-b668-ddcc38a6b26f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:08:35Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 00:08:05.013265 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 00:08:05.015776 1 observer_polling.go:159] Starting file observer\\\\nI0318 00:08:05.046728 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 00:08:05.051577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 00:08:35.404406 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:13Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.086204 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:13Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.105428 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:13Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.125141 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:13Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.141284 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:13Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.159715 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:13Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.184616 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:13Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.203799 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:13Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.844123 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:13 crc kubenswrapper[4983]: E0318 00:10:13.844520 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.844810 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:13 crc kubenswrapper[4983]: E0318 00:10:13.844914 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.845036 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:13 crc kubenswrapper[4983]: I0318 00:10:13.845146 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:13 crc kubenswrapper[4983]: E0318 00:10:13.845239 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:13 crc kubenswrapper[4983]: E0318 00:10:13.845344 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:15 crc kubenswrapper[4983]: I0318 00:10:15.843919 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:15 crc kubenswrapper[4983]: I0318 00:10:15.844012 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:15 crc kubenswrapper[4983]: I0318 00:10:15.844039 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:15 crc kubenswrapper[4983]: E0318 00:10:15.844131 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:15 crc kubenswrapper[4983]: I0318 00:10:15.843962 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:15 crc kubenswrapper[4983]: E0318 00:10:15.844415 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:15 crc kubenswrapper[4983]: E0318 00:10:15.844485 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:15 crc kubenswrapper[4983]: E0318 00:10:15.844622 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:15 crc kubenswrapper[4983]: I0318 00:10:15.859367 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 18 00:10:17 crc kubenswrapper[4983]: I0318 00:10:17.843575 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:17 crc kubenswrapper[4983]: I0318 00:10:17.843628 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:17 crc kubenswrapper[4983]: I0318 00:10:17.843662 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:17 crc kubenswrapper[4983]: I0318 00:10:17.843648 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:17 crc kubenswrapper[4983]: E0318 00:10:17.843779 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:17 crc kubenswrapper[4983]: E0318 00:10:17.843910 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:17 crc kubenswrapper[4983]: E0318 00:10:17.844074 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:17 crc kubenswrapper[4983]: E0318 00:10:17.844257 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:17 crc kubenswrapper[4983]: E0318 00:10:17.963527 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:10:19 crc kubenswrapper[4983]: I0318 00:10:19.844156 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:19 crc kubenswrapper[4983]: I0318 00:10:19.844205 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:19 crc kubenswrapper[4983]: I0318 00:10:19.844305 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:19 crc kubenswrapper[4983]: E0318 00:10:19.844338 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:19 crc kubenswrapper[4983]: I0318 00:10:19.844181 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:19 crc kubenswrapper[4983]: E0318 00:10:19.844472 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:19 crc kubenswrapper[4983]: E0318 00:10:19.844625 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:19 crc kubenswrapper[4983]: E0318 00:10:19.844740 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.896257 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.896329 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.896352 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.896380 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.896402 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:20Z","lastTransitionTime":"2026-03-18T00:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:20 crc kubenswrapper[4983]: E0318 00:10:20.920085 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:20Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.925660 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.925833 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.925887 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.925918 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.925959 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:20Z","lastTransitionTime":"2026-03-18T00:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:20 crc kubenswrapper[4983]: E0318 00:10:20.948529 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:20Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.953998 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.954044 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.954062 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.954087 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.954106 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:20Z","lastTransitionTime":"2026-03-18T00:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:20 crc kubenswrapper[4983]: E0318 00:10:20.975005 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:20Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.979806 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.979905 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.979924 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.980008 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:20 crc kubenswrapper[4983]: I0318 00:10:20.980028 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:20Z","lastTransitionTime":"2026-03-18T00:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:20 crc kubenswrapper[4983]: E0318 00:10:20.998733 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:20Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:21 crc kubenswrapper[4983]: I0318 00:10:21.004501 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:21 crc kubenswrapper[4983]: I0318 00:10:21.004548 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:21 crc kubenswrapper[4983]: I0318 00:10:21.004570 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:21 crc kubenswrapper[4983]: I0318 00:10:21.004599 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:21 crc kubenswrapper[4983]: I0318 00:10:21.004617 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:21Z","lastTransitionTime":"2026-03-18T00:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:21 crc kubenswrapper[4983]: E0318 00:10:21.025863 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:21Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:21 crc kubenswrapper[4983]: E0318 00:10:21.026140 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 00:10:21 crc kubenswrapper[4983]: I0318 00:10:21.844205 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:21 crc kubenswrapper[4983]: I0318 00:10:21.844284 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:21 crc kubenswrapper[4983]: I0318 00:10:21.844237 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:21 crc kubenswrapper[4983]: I0318 00:10:21.844205 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:21 crc kubenswrapper[4983]: E0318 00:10:21.844421 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:21 crc kubenswrapper[4983]: E0318 00:10:21.844581 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:21 crc kubenswrapper[4983]: E0318 00:10:21.844805 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:21 crc kubenswrapper[4983]: E0318 00:10:21.844864 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:22 crc kubenswrapper[4983]: I0318 00:10:22.864147 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:22Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:22 crc kubenswrapper[4983]: I0318 00:10:22.886434 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:22Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:22 crc kubenswrapper[4983]: I0318 00:10:22.907867 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:22Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:22 crc kubenswrapper[4983]: I0318 00:10:22.928087 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e68e246-bb9b-4b78-b668-ddcc38a6b26f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:08:35Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 00:08:05.013265 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 00:08:05.015776 1 observer_polling.go:159] Starting file observer\\\\nI0318 00:08:05.046728 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 00:08:05.051577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 00:08:35.404406 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:22Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:22 crc kubenswrapper[4983]: I0318 00:10:22.948279 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:22Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:22 crc kubenswrapper[4983]: E0318 00:10:22.964443 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:10:22 crc kubenswrapper[4983]: I0318 00:10:22.969041 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:22Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:22 crc kubenswrapper[4983]: I0318 00:10:22.990145 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:22Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.009793 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:23Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.041637 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:23Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.056183 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:23Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.074152 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:23Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.088527 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:23Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.104973 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:23Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.120907 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cade86f-9f25-4f96-b861-8857ce2ad162\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81167a1f7781d69d700f7647928cbb027a40efec3129a860f62f82b12e2d9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84f4ccd83826ef23c1583cae04ee8092f94bb5a2f9c948f97e503a9b1858201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3939424d6655bea60e9bad967f755a313411e8a699563cd3a9ef2b2e2354362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:23Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.139709 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:23Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.158586 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:23Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.195242 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:08Z\\\",\\\"message\\\":\\\"rts Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 00:10:08.902818 7235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}\\\\nF0318 00:10:08.903148 7235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:23Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.215571 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:23Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.844249 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.844358 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.844294 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:23 crc kubenswrapper[4983]: E0318 00:10:23.844478 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:23 crc kubenswrapper[4983]: E0318 00:10:23.844615 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:23 crc kubenswrapper[4983]: E0318 00:10:23.844713 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:23 crc kubenswrapper[4983]: I0318 00:10:23.845072 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:23 crc kubenswrapper[4983]: E0318 00:10:23.845174 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:24 crc kubenswrapper[4983]: I0318 00:10:24.845381 4983 scope.go:117] "RemoveContainer" containerID="89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707" Mar 18 00:10:24 crc kubenswrapper[4983]: E0318 00:10:24.845729 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" Mar 18 00:10:25 crc kubenswrapper[4983]: I0318 00:10:25.844209 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:25 crc kubenswrapper[4983]: I0318 00:10:25.844321 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:25 crc kubenswrapper[4983]: I0318 00:10:25.844320 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:25 crc kubenswrapper[4983]: E0318 00:10:25.844387 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:25 crc kubenswrapper[4983]: I0318 00:10:25.844233 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:25 crc kubenswrapper[4983]: E0318 00:10:25.844547 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:25 crc kubenswrapper[4983]: E0318 00:10:25.844604 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:25 crc kubenswrapper[4983]: E0318 00:10:25.844732 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:27 crc kubenswrapper[4983]: I0318 00:10:27.844054 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:27 crc kubenswrapper[4983]: I0318 00:10:27.844076 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:27 crc kubenswrapper[4983]: E0318 00:10:27.844702 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:27 crc kubenswrapper[4983]: I0318 00:10:27.844173 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:27 crc kubenswrapper[4983]: I0318 00:10:27.844124 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:27 crc kubenswrapper[4983]: E0318 00:10:27.844839 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:27 crc kubenswrapper[4983]: E0318 00:10:27.845029 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:27 crc kubenswrapper[4983]: E0318 00:10:27.845205 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:27 crc kubenswrapper[4983]: E0318 00:10:27.965795 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:10:28 crc kubenswrapper[4983]: I0318 00:10:28.547552 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:28 crc kubenswrapper[4983]: E0318 00:10:28.547811 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:10:28 crc kubenswrapper[4983]: E0318 00:10:28.547995 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs podName:fe22dcea-cd03-43e3-80da-61a5f1e063ad nodeName:}" failed. No retries permitted until 2026-03-18 00:11:00.547922897 +0000 UTC m=+178.445297557 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs") pod "network-metrics-daemon-vjcg5" (UID: "fe22dcea-cd03-43e3-80da-61a5f1e063ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:10:29 crc kubenswrapper[4983]: I0318 00:10:29.843963 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:29 crc kubenswrapper[4983]: I0318 00:10:29.843962 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:29 crc kubenswrapper[4983]: I0318 00:10:29.844093 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:29 crc kubenswrapper[4983]: E0318 00:10:29.845304 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:29 crc kubenswrapper[4983]: E0318 00:10:29.845391 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:29 crc kubenswrapper[4983]: I0318 00:10:29.844172 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:29 crc kubenswrapper[4983]: E0318 00:10:29.846229 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:29 crc kubenswrapper[4983]: E0318 00:10:29.846362 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.606402 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6nblt_0853f8de-9bfc-4053-acd0-8893122d2d62/kube-multus/0.log" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.606465 4983 generic.go:334] "Generic (PLEG): container finished" podID="0853f8de-9bfc-4053-acd0-8893122d2d62" containerID="ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2" exitCode=1 Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.606505 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6nblt" event={"ID":"0853f8de-9bfc-4053-acd0-8893122d2d62","Type":"ContainerDied","Data":"ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2"} Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.607188 4983 scope.go:117] "RemoveContainer" containerID="ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.629910 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e68e246-bb9b-4b78-b668-ddcc38a6b26f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:08:35Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 00:08:05.013265 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 00:08:05.015776 1 observer_polling.go:159] Starting file observer\\\\nI0318 00:08:05.046728 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 00:08:05.051577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 00:08:35.404406 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.652284 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.672542 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.692239 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.712556 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.731019 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.757721 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.779639 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.795114 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.817069 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:30Z\\\",\\\"message\\\":\\\"2026-03-18T00:09:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c\\\\n2026-03-18T00:09:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c to /host/opt/cni/bin/\\\\n2026-03-18T00:09:45Z [verbose] multus-daemon started\\\\n2026-03-18T00:09:45Z [verbose] Readiness Indicator file check\\\\n2026-03-18T00:10:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.831387 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.864686 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.881260 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.900431 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.917750 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.951863 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:08Z\\\",\\\"message\\\":\\\"rts Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 00:10:08.902818 7235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}\\\\nF0318 00:10:08.903148 7235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.967570 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:30 crc kubenswrapper[4983]: I0318 00:10:30.984291 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cade86f-9f25-4f96-b861-8857ce2ad162\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81167a1f7781d69d700f7647928cbb027a40efec3129a860f62f82b12e2d9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84f4ccd83826ef23c1583cae04ee8092f94bb5a2f9c948f97e503a9b1858201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3939424d6655bea60e9bad967f755a313411e8a699563cd3a9ef2b2e2354362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:30Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.036957 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.037026 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.037050 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.037081 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.037103 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:31Z","lastTransitionTime":"2026-03-18T00:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.058870 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.063859 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.063958 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.063974 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.063991 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.064003 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:31Z","lastTransitionTime":"2026-03-18T00:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.088366 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.092920 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.093004 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.093021 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.093044 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.093061 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:31Z","lastTransitionTime":"2026-03-18T00:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.112806 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.117244 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.117306 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.117323 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.117348 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.117365 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:31Z","lastTransitionTime":"2026-03-18T00:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.134414 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.138914 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.139008 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.139065 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.139097 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.139123 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:31Z","lastTransitionTime":"2026-03-18T00:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.158527 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.158755 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.612847 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6nblt_0853f8de-9bfc-4053-acd0-8893122d2d62/kube-multus/0.log" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.612907 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6nblt" event={"ID":"0853f8de-9bfc-4053-acd0-8893122d2d62","Type":"ContainerStarted","Data":"96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b"} Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.626637 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.642604 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:30Z\\\",\\\"message\\\":\\\"2026-03-18T00:09:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c\\\\n2026-03-18T00:09:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c to /host/opt/cni/bin/\\\\n2026-03-18T00:09:45Z [verbose] multus-daemon started\\\\n2026-03-18T00:09:45Z [verbose] Readiness Indicator file check\\\\n2026-03-18T00:10:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.655208 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.678894 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.692346 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.708319 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.727255 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.749165 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:08Z\\\",\\\"message\\\":\\\"rts Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 00:10:08.902818 7235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}\\\\nF0318 00:10:08.903148 7235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.762813 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.778253 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cade86f-9f25-4f96-b861-8857ce2ad162\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81167a1f7781d69d700f7647928cbb027a40efec3129a860f62f82b12e2d9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84f4ccd83826ef23c1583cae04ee8092f94bb5a2f9c948f97e503a9b1858201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3939424d6655bea60e9bad967f755a313411e8a699563cd3a9ef2b2e2354362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.783093 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.783230 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.783245 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:35.783217914 +0000 UTC m=+213.680592564 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.783288 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.783332 4983 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.783394 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:11:35.783377409 +0000 UTC m=+213.680752049 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.783336 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.783445 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.783510 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.783550 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.783570 4983 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.783577 4983 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.783496 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.783632 4983 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.783643 4983 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.783614 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 00:11:35.783603577 +0000 UTC m=+213.680978187 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.783696 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 00:11:35.783677179 +0000 UTC m=+213.681051829 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.783719 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 00:11:35.78370753 +0000 UTC m=+213.681082180 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.795253 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e68e246-bb9b-4b78-b668-ddcc38a6b26f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:08:35Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 00:08:05.013265 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 00:08:05.015776 1 observer_polling.go:159] Starting file observer\\\\nI0318 00:08:05.046728 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 00:08:05.051577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 00:08:35.404406 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.810446 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.823299 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.836736 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.843675 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.843714 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.843781 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.843779 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.844068 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.844091 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.844167 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:31 crc kubenswrapper[4983]: E0318 00:10:31.844329 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.848415 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.860414 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.875306 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:31 crc kubenswrapper[4983]: I0318 00:10:31.894435 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:31Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:32 crc kubenswrapper[4983]: I0318 00:10:32.867349 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:32Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:32 crc kubenswrapper[4983]: I0318 00:10:32.884636 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:32Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:32 crc kubenswrapper[4983]: I0318 00:10:32.900412 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:32Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:32 crc kubenswrapper[4983]: I0318 00:10:32.920618 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:32Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:32 crc kubenswrapper[4983]: I0318 00:10:32.935036 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:32Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:32 crc kubenswrapper[4983]: I0318 00:10:32.953832 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e68e246-bb9b-4b78-b668-ddcc38a6b26f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:08:35Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 00:08:05.013265 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 00:08:05.015776 1 observer_polling.go:159] Starting file observer\\\\nI0318 00:08:05.046728 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 00:08:05.051577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 00:08:35.404406 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:32Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:32 crc kubenswrapper[4983]: E0318 00:10:32.966714 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:10:32 crc kubenswrapper[4983]: I0318 00:10:32.969472 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:32Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:32 crc kubenswrapper[4983]: I0318 00:10:32.987833 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:32Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.003748 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:33Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.031430 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:33Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.049444 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:33Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.072011 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:30Z\\\",\\\"message\\\":\\\"2026-03-18T00:09:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c\\\\n2026-03-18T00:09:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c to /host/opt/cni/bin/\\\\n2026-03-18T00:09:45Z [verbose] multus-daemon started\\\\n2026-03-18T00:09:45Z [verbose] Readiness Indicator file check\\\\n2026-03-18T00:10:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:33Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.087888 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:33Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.101511 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:33Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.117951 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cade86f-9f25-4f96-b861-8857ce2ad162\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81167a1f7781d69d700f7647928cbb027a40efec3129a860f62f82b12e2d9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84f4ccd83826ef23c1583cae04ee8092f94bb5a2f9c948f97e503a9b1858201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3939424d6655bea60e9bad967f755a313411e8a699563cd3a9ef2b2e2354362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:33Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.166913 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:33Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.180105 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:33Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.211463 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:08Z\\\",\\\"message\\\":\\\"rts Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 00:10:08.902818 7235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}\\\\nF0318 00:10:08.903148 7235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:33Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.844625 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.844697 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.844662 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:33 crc kubenswrapper[4983]: I0318 00:10:33.844646 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:33 crc kubenswrapper[4983]: E0318 00:10:33.844850 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:33 crc kubenswrapper[4983]: E0318 00:10:33.844956 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:33 crc kubenswrapper[4983]: E0318 00:10:33.845017 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:33 crc kubenswrapper[4983]: E0318 00:10:33.845096 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:35 crc kubenswrapper[4983]: I0318 00:10:35.844106 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:35 crc kubenswrapper[4983]: I0318 00:10:35.844190 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:35 crc kubenswrapper[4983]: I0318 00:10:35.844279 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:35 crc kubenswrapper[4983]: E0318 00:10:35.844720 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:35 crc kubenswrapper[4983]: E0318 00:10:35.844542 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:35 crc kubenswrapper[4983]: E0318 00:10:35.844824 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:35 crc kubenswrapper[4983]: I0318 00:10:35.844290 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:35 crc kubenswrapper[4983]: E0318 00:10:35.844992 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:37 crc kubenswrapper[4983]: I0318 00:10:37.843975 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:37 crc kubenswrapper[4983]: I0318 00:10:37.844044 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:37 crc kubenswrapper[4983]: E0318 00:10:37.844200 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:37 crc kubenswrapper[4983]: I0318 00:10:37.844308 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:37 crc kubenswrapper[4983]: E0318 00:10:37.844360 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:37 crc kubenswrapper[4983]: E0318 00:10:37.844554 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:37 crc kubenswrapper[4983]: I0318 00:10:37.845449 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:37 crc kubenswrapper[4983]: E0318 00:10:37.845717 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:37 crc kubenswrapper[4983]: E0318 00:10:37.968566 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:10:39 crc kubenswrapper[4983]: I0318 00:10:39.844198 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:39 crc kubenswrapper[4983]: I0318 00:10:39.844268 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:39 crc kubenswrapper[4983]: E0318 00:10:39.845550 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:39 crc kubenswrapper[4983]: I0318 00:10:39.845016 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:39 crc kubenswrapper[4983]: I0318 00:10:39.845627 4983 scope.go:117] "RemoveContainer" containerID="89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707" Mar 18 00:10:39 crc kubenswrapper[4983]: I0318 00:10:39.844314 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:39 crc kubenswrapper[4983]: E0318 00:10:39.845735 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:39 crc kubenswrapper[4983]: E0318 00:10:39.845656 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:39 crc kubenswrapper[4983]: E0318 00:10:39.845842 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.651558 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/2.log" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.656582 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerStarted","Data":"2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e"} Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.657294 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.680818 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.715887 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:08Z\\\",\\\"message\\\":\\\"rts Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 00:10:08.902818 7235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}\\\\nF0318 00:10:08.903148 7235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:10:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.736386 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.755480 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cade86f-9f25-4f96-b861-8857ce2ad162\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81167a1f7781d69d700f7647928cbb027a40efec3129a860f62f82b12e2d9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84f4ccd83826ef23c1583cae04ee8092f94bb5a2f9c948f97e503a9b1858201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3939424d6655bea60e9bad967f755a313411e8a699563cd3a9ef2b2e2354362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.790314 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.814149 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.838677 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.864458 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.888573 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.909515 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.933526 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.951692 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.969335 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e68e246-bb9b-4b78-b668-ddcc38a6b26f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:08:35Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 00:08:05.013265 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 00:08:05.015776 1 observer_polling.go:159] Starting file observer\\\\nI0318 00:08:05.046728 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 00:08:05.051577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 00:08:35.404406 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.980678 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:40 crc kubenswrapper[4983]: I0318 00:10:40.995440 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:30Z\\\",\\\"message\\\":\\\"2026-03-18T00:09:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c\\\\n2026-03-18T00:09:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c to /host/opt/cni/bin/\\\\n2026-03-18T00:09:45Z [verbose] multus-daemon started\\\\n2026-03-18T00:09:45Z [verbose] Readiness Indicator file check\\\\n2026-03-18T00:10:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:40Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.010106 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.044305 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.063522 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.204019 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.204076 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.204101 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.204129 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.204153 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:41Z","lastTransitionTime":"2026-03-18T00:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:41 crc kubenswrapper[4983]: E0318 00:10:41.221899 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.226540 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.226677 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.226762 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.226860 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.226961 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:41Z","lastTransitionTime":"2026-03-18T00:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:41 crc kubenswrapper[4983]: E0318 00:10:41.244686 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.249461 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.249529 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.249549 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.249576 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.249599 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:41Z","lastTransitionTime":"2026-03-18T00:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:41 crc kubenswrapper[4983]: E0318 00:10:41.270821 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.275256 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.275323 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.275342 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.275364 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.275384 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:41Z","lastTransitionTime":"2026-03-18T00:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:41 crc kubenswrapper[4983]: E0318 00:10:41.295463 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.300297 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.300355 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.300380 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.300407 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.300430 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:41Z","lastTransitionTime":"2026-03-18T00:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:41 crc kubenswrapper[4983]: E0318 00:10:41.320667 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: E0318 00:10:41.321003 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.664614 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/3.log" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.666118 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/2.log" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.670484 4983 generic.go:334] "Generic (PLEG): container finished" podID="938043fe-450c-4477-8f55-312a843b0706" containerID="2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e" exitCode=1 Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.670540 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e"} Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.670595 4983 scope.go:117] "RemoveContainer" containerID="89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.671521 4983 scope.go:117] "RemoveContainer" containerID="2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e" Mar 18 00:10:41 crc kubenswrapper[4983]: E0318 00:10:41.671771 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.693860 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.713700 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.740524 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.760916 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.778139 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.801111 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.821725 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.841758 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e68e246-bb9b-4b78-b668-ddcc38a6b26f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:08:35Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 00:08:05.013265 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 00:08:05.015776 1 observer_polling.go:159] Starting file observer\\\\nI0318 00:08:05.046728 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 00:08:05.051577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 00:08:35.404406 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.844013 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.844082 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.844157 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:41 crc kubenswrapper[4983]: E0318 00:10:41.844359 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.844386 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:41 crc kubenswrapper[4983]: E0318 00:10:41.844547 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:41 crc kubenswrapper[4983]: E0318 00:10:41.844627 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:41 crc kubenswrapper[4983]: E0318 00:10:41.844744 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.857917 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.879392 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:30Z\\\",\\\"message\\\":\\\"2026-03-18T00:09:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c\\\\n2026-03-18T00:09:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c to /host/opt/cni/bin/\\\\n2026-03-18T00:09:45Z [verbose] multus-daemon started\\\\n2026-03-18T00:09:45Z [verbose] Readiness Indicator file check\\\\n2026-03-18T00:10:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.895319 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.927812 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.946324 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:41 crc kubenswrapper[4983]: I0318 00:10:41.967631 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.000354 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89c0938d168630ee638db0cf83707c6c4e54126dbaf06081d07ea075cfc7c707\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:08Z\\\",\\\"message\\\":\\\"rts Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 00:10:08.902818 7235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}\\\\nF0318 00:10:08.903148 7235 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:40Z\\\",\\\"message\\\":\\\"7 7553 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889001 7553 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889149 7553 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889194 7553 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889258 7553 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:10:40.889459 7553 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0318 00:10:40.890665 7553 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:41Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.017504 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.035065 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cade86f-9f25-4f96-b861-8857ce2ad162\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81167a1f7781d69d700f7647928cbb027a40efec3129a860f62f82b12e2d9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84f4ccd83826ef23c1583cae04ee8092f94bb5a2f9c948f97e503a9b1858201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3939424d6655bea60e9bad967f755a313411e8a699563cd3a9ef2b2e2354362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.058282 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.677498 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/3.log" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.683175 4983 scope.go:117] "RemoveContainer" containerID="2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e" Mar 18 00:10:42 crc kubenswrapper[4983]: E0318 00:10:42.683452 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.702847 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.734696 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:40Z\\\",\\\"message\\\":\\\"7 7553 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889001 7553 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889149 7553 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889194 7553 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889258 7553 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:10:40.889459 7553 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0318 00:10:40.890665 7553 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.753061 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.771975 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cade86f-9f25-4f96-b861-8857ce2ad162\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81167a1f7781d69d700f7647928cbb027a40efec3129a860f62f82b12e2d9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84f4ccd83826ef23c1583cae04ee8092f94bb5a2f9c948f97e503a9b1858201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3939424d6655bea60e9bad967f755a313411e8a699563cd3a9ef2b2e2354362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.792170 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.813411 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.832125 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.853500 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.872447 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.890012 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.915466 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.937728 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.959182 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e68e246-bb9b-4b78-b668-ddcc38a6b26f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:08:35Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 00:08:05.013265 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 00:08:05.015776 1 observer_polling.go:159] Starting file observer\\\\nI0318 00:08:05.046728 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 00:08:05.051577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 00:08:35.404406 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:42 crc kubenswrapper[4983]: E0318 00:10:42.969690 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:10:42 crc kubenswrapper[4983]: I0318 00:10:42.976830 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.002077 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:30Z\\\",\\\"message\\\":\\\"2026-03-18T00:09:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c\\\\n2026-03-18T00:09:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c to /host/opt/cni/bin/\\\\n2026-03-18T00:09:45Z [verbose] multus-daemon started\\\\n2026-03-18T00:09:45Z [verbose] Readiness Indicator file check\\\\n2026-03-18T00:10:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:42Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.018004 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.050260 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.067486 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.084981 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.102252 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.129164 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.151514 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.171976 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e68e246-bb9b-4b78-b668-ddcc38a6b26f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:08:35Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 00:08:05.013265 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 00:08:05.015776 1 observer_polling.go:159] Starting file observer\\\\nI0318 00:08:05.046728 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 00:08:05.051577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 00:08:35.404406 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.192868 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.211759 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.231162 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.265968 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.282357 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.302537 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:30Z\\\",\\\"message\\\":\\\"2026-03-18T00:09:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c\\\\n2026-03-18T00:09:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c to /host/opt/cni/bin/\\\\n2026-03-18T00:09:45Z [verbose] multus-daemon started\\\\n2026-03-18T00:09:45Z [verbose] Readiness Indicator file check\\\\n2026-03-18T00:10:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.318921 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.334882 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.353760 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cade86f-9f25-4f96-b861-8857ce2ad162\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81167a1f7781d69d700f7647928cbb027a40efec3129a860f62f82b12e2d9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84f4ccd83826ef23c1583cae04ee8092f94bb5a2f9c948f97e503a9b1858201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3939424d6655bea60e9bad967f755a313411e8a699563cd3a9ef2b2e2354362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.380124 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.400010 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.433137 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:40Z\\\",\\\"message\\\":\\\"7 7553 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889001 7553 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889149 7553 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889194 7553 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889258 7553 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:10:40.889459 7553 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0318 00:10:40.890665 7553 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.455342 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:43Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.844441 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.844501 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:43 crc kubenswrapper[4983]: E0318 00:10:43.844674 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.844733 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:43 crc kubenswrapper[4983]: I0318 00:10:43.844771 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:43 crc kubenswrapper[4983]: E0318 00:10:43.844901 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:43 crc kubenswrapper[4983]: E0318 00:10:43.845113 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:43 crc kubenswrapper[4983]: E0318 00:10:43.845243 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:44 crc kubenswrapper[4983]: I0318 00:10:44.863905 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 18 00:10:45 crc kubenswrapper[4983]: I0318 00:10:45.843422 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:45 crc kubenswrapper[4983]: I0318 00:10:45.843433 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:45 crc kubenswrapper[4983]: E0318 00:10:45.843897 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:45 crc kubenswrapper[4983]: I0318 00:10:45.843557 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:45 crc kubenswrapper[4983]: I0318 00:10:45.843487 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:45 crc kubenswrapper[4983]: E0318 00:10:45.844650 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:45 crc kubenswrapper[4983]: E0318 00:10:45.844109 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:45 crc kubenswrapper[4983]: E0318 00:10:45.845227 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:47 crc kubenswrapper[4983]: I0318 00:10:47.844533 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:47 crc kubenswrapper[4983]: I0318 00:10:47.844586 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:47 crc kubenswrapper[4983]: E0318 00:10:47.845082 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:47 crc kubenswrapper[4983]: I0318 00:10:47.844724 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:47 crc kubenswrapper[4983]: I0318 00:10:47.844639 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:47 crc kubenswrapper[4983]: E0318 00:10:47.845239 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:47 crc kubenswrapper[4983]: E0318 00:10:47.845398 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:47 crc kubenswrapper[4983]: E0318 00:10:47.845622 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:47 crc kubenswrapper[4983]: E0318 00:10:47.971073 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:10:49 crc kubenswrapper[4983]: I0318 00:10:49.843853 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:49 crc kubenswrapper[4983]: I0318 00:10:49.843917 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:49 crc kubenswrapper[4983]: I0318 00:10:49.844054 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:49 crc kubenswrapper[4983]: I0318 00:10:49.843870 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:49 crc kubenswrapper[4983]: E0318 00:10:49.844187 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:49 crc kubenswrapper[4983]: E0318 00:10:49.844370 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:49 crc kubenswrapper[4983]: E0318 00:10:49.844505 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:49 crc kubenswrapper[4983]: E0318 00:10:49.844722 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.621642 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.621709 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.621726 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.621751 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.621767 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:51Z","lastTransitionTime":"2026-03-18T00:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:51 crc kubenswrapper[4983]: E0318 00:10:51.646269 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.650967 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.651008 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.651018 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.651037 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.651048 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:51Z","lastTransitionTime":"2026-03-18T00:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:51 crc kubenswrapper[4983]: E0318 00:10:51.670962 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.676184 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.676241 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.676259 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.676286 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.676304 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:51Z","lastTransitionTime":"2026-03-18T00:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:51 crc kubenswrapper[4983]: E0318 00:10:51.696160 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.705493 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.705529 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.705538 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.705552 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.705562 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:51Z","lastTransitionTime":"2026-03-18T00:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:51 crc kubenswrapper[4983]: E0318 00:10:51.725634 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.729154 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.729218 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.729239 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.729264 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.729283 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:10:51Z","lastTransitionTime":"2026-03-18T00:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:10:51 crc kubenswrapper[4983]: E0318 00:10:51.746575 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:51Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:51 crc kubenswrapper[4983]: E0318 00:10:51.746679 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.843703 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:51 crc kubenswrapper[4983]: E0318 00:10:51.843830 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.844053 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:51 crc kubenswrapper[4983]: E0318 00:10:51.844119 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.844223 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:51 crc kubenswrapper[4983]: I0318 00:10:51.844338 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:51 crc kubenswrapper[4983]: E0318 00:10:51.844358 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:51 crc kubenswrapper[4983]: E0318 00:10:51.844577 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:52 crc kubenswrapper[4983]: I0318 00:10:52.859665 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:52Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:52 crc kubenswrapper[4983]: I0318 00:10:52.877371 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cade86f-9f25-4f96-b861-8857ce2ad162\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81167a1f7781d69d700f7647928cbb027a40efec3129a860f62f82b12e2d9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84f4ccd83826ef23c1583cae04ee8092f94bb5a2f9c948f97e503a9b1858201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3939424d6655bea60e9bad967f755a313411e8a699563cd3a9ef2b2e2354362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:52Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:52 crc kubenswrapper[4983]: I0318 00:10:52.896905 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:52Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:52 crc kubenswrapper[4983]: I0318 00:10:52.915336 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:52Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:52 crc kubenswrapper[4983]: I0318 00:10:52.945169 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:40Z\\\",\\\"message\\\":\\\"7 7553 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889001 7553 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889149 7553 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889194 7553 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889258 7553 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:10:40.889459 7553 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0318 00:10:40.890665 7553 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:52Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:52 crc kubenswrapper[4983]: I0318 00:10:52.964647 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:52Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:52 crc kubenswrapper[4983]: E0318 00:10:52.971910 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:10:52 crc kubenswrapper[4983]: I0318 00:10:52.986888 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:52Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.003836 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.024909 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.045168 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.064245 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e68e246-bb9b-4b78-b668-ddcc38a6b26f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:08:35Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 00:08:05.013265 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 00:08:05.015776 1 observer_polling.go:159] Starting file observer\\\\nI0318 00:08:05.046728 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 00:08:05.051577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 00:08:35.404406 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.083410 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.101688 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.117372 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.149812 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.165980 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19860e89-0f4a-4c9f-b616-8329f0b9bab8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0befc5603e720538a3a0e9c40a50792f3e8d07c61929b87bd424427be0e54ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655e9f142d07c3af06937170bb4eaa3ec5fbbb608685854cc1545665aec22b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://655e9f142d07c3af06937170bb4eaa3ec5fbbb608685854cc1545665aec22b24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.182857 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.203893 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:30Z\\\",\\\"message\\\":\\\"2026-03-18T00:09:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c\\\\n2026-03-18T00:09:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c to /host/opt/cni/bin/\\\\n2026-03-18T00:09:45Z [verbose] multus-daemon started\\\\n2026-03-18T00:09:45Z [verbose] Readiness Indicator file check\\\\n2026-03-18T00:10:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.218459 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:10:53Z is after 2025-08-24T17:21:41Z" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.843922 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.843989 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.844108 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:53 crc kubenswrapper[4983]: I0318 00:10:53.844178 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:53 crc kubenswrapper[4983]: E0318 00:10:53.844144 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:53 crc kubenswrapper[4983]: E0318 00:10:53.844317 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:53 crc kubenswrapper[4983]: E0318 00:10:53.844446 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:53 crc kubenswrapper[4983]: E0318 00:10:53.844643 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:54 crc kubenswrapper[4983]: I0318 00:10:54.845767 4983 scope.go:117] "RemoveContainer" containerID="2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e" Mar 18 00:10:54 crc kubenswrapper[4983]: E0318 00:10:54.846429 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" Mar 18 00:10:55 crc kubenswrapper[4983]: I0318 00:10:55.843470 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:55 crc kubenswrapper[4983]: I0318 00:10:55.843587 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:55 crc kubenswrapper[4983]: I0318 00:10:55.843550 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:55 crc kubenswrapper[4983]: E0318 00:10:55.843679 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:55 crc kubenswrapper[4983]: I0318 00:10:55.843723 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:55 crc kubenswrapper[4983]: E0318 00:10:55.843768 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:55 crc kubenswrapper[4983]: E0318 00:10:55.843900 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:55 crc kubenswrapper[4983]: E0318 00:10:55.844057 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:57 crc kubenswrapper[4983]: I0318 00:10:57.843750 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:57 crc kubenswrapper[4983]: E0318 00:10:57.843918 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:57 crc kubenswrapper[4983]: I0318 00:10:57.844094 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:57 crc kubenswrapper[4983]: I0318 00:10:57.844183 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:57 crc kubenswrapper[4983]: E0318 00:10:57.844273 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:57 crc kubenswrapper[4983]: E0318 00:10:57.844379 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:10:57 crc kubenswrapper[4983]: I0318 00:10:57.844475 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:57 crc kubenswrapper[4983]: E0318 00:10:57.844623 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:57 crc kubenswrapper[4983]: E0318 00:10:57.973873 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:10:59 crc kubenswrapper[4983]: I0318 00:10:59.843466 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:10:59 crc kubenswrapper[4983]: I0318 00:10:59.843581 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:10:59 crc kubenswrapper[4983]: I0318 00:10:59.843710 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:10:59 crc kubenswrapper[4983]: E0318 00:10:59.843838 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:10:59 crc kubenswrapper[4983]: I0318 00:10:59.843883 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:10:59 crc kubenswrapper[4983]: E0318 00:10:59.844080 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:10:59 crc kubenswrapper[4983]: E0318 00:10:59.844212 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:10:59 crc kubenswrapper[4983]: E0318 00:10:59.844321 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:00 crc kubenswrapper[4983]: I0318 00:11:00.620241 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:00 crc kubenswrapper[4983]: E0318 00:11:00.620694 4983 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:11:00 crc kubenswrapper[4983]: E0318 00:11:00.620832 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs podName:fe22dcea-cd03-43e3-80da-61a5f1e063ad nodeName:}" failed. No retries permitted until 2026-03-18 00:12:04.62078439 +0000 UTC m=+242.518159070 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs") pod "network-metrics-daemon-vjcg5" (UID: "fe22dcea-cd03-43e3-80da-61a5f1e063ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 00:11:01 crc kubenswrapper[4983]: I0318 00:11:01.988546 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:01 crc kubenswrapper[4983]: E0318 00:11:01.988695 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:01 crc kubenswrapper[4983]: I0318 00:11:01.988964 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:01 crc kubenswrapper[4983]: I0318 00:11:01.988998 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:01 crc kubenswrapper[4983]: E0318 00:11:01.989062 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:01 crc kubenswrapper[4983]: E0318 00:11:01.989257 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:01 crc kubenswrapper[4983]: I0318 00:11:01.989888 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:01 crc kubenswrapper[4983]: E0318 00:11:01.990105 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.052101 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.052139 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.052150 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.052169 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.052192 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:11:02Z","lastTransitionTime":"2026-03-18T00:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:11:02 crc kubenswrapper[4983]: E0318 00:11:02.065711 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.070193 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.070404 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.070659 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.070840 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.071034 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:11:02Z","lastTransitionTime":"2026-03-18T00:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:11:02 crc kubenswrapper[4983]: E0318 00:11:02.085161 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.090423 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.090598 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.090764 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.090894 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.091062 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:11:02Z","lastTransitionTime":"2026-03-18T00:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:11:02 crc kubenswrapper[4983]: E0318 00:11:02.107965 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.112033 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.112228 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.112256 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.112316 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.112495 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:11:02Z","lastTransitionTime":"2026-03-18T00:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:11:02 crc kubenswrapper[4983]: E0318 00:11:02.126689 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.130630 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.130669 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.130683 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.130704 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.130718 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:11:02Z","lastTransitionTime":"2026-03-18T00:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:11:02 crc kubenswrapper[4983]: E0318 00:11:02.144749 4983 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T00:11:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"20e16f4a-2b2c-47e6-b1f1-f0e9e09bc4c5\\\",\\\"systemUUID\\\":\\\"3cf496ed-375e-4961-a03c-4bc8923d9044\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:02 crc kubenswrapper[4983]: E0318 00:11:02.145169 4983 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.867443 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bf7769db3e60332004ed61aab8cbaff2a5e2082b6e2c86815abaeb941b6a0e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.889682 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"482e2afe-48be-4ea2-8bda-52849c901b52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c93f884c1b99e8efbfcdc9b67ffd473f8399a931ab3eedeedae9eeb42ead8bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kfjxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jczfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.917201 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-69n75" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72a1209f-2f3a-48e3-ade2-7aa5824f45eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25737247fa6decbbd8c0305f2ae5cbe892658fd1464376158e7a480d7b26a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c191b41d5dfa25ec621a7be5d93d6bc625d06022e4a2f420d49c41fde136ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://730c31d047e6e2766d638cb288b6570713c376382fd15768b93ad5b62bdc2e2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1deb707d1748cb4c5bd2ccfba5db5db1654ed2a1ef1d48a95c31fde4eca09beb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9148b31a3018236124f0cffcc81231f5002249430840a14cd082c64ec9f08d9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4083cc59ddc56bb31a0c91989e1fde60ef976bbfad7f80b28d6f9d32c3f3028f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e031cad2d9e089f8f4afa7b8b6dd13e24572ef490dd362ba7cf7712431859a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g8459\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-69n75\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.940538 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f49d12d-9e58-4eae-a24d-b58a49f81e29\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:09:04Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 00:09:03.623256 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 00:09:03.623355 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 00:09:03.623852 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3798277013/tls.crt::/tmp/serving-cert-3798277013/tls.key\\\\\\\"\\\\nI0318 00:09:03.975037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 00:09:03.979528 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 00:09:03.979550 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 00:09:03.979576 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 00:09:03.979583 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 00:09:03.988357 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0318 00:09:03.988376 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0318 00:09:03.988422 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988434 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 00:09:03.988446 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 00:09:03.988455 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 00:09:03.988466 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 00:09:03.988474 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0318 00:09:03.990245 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.962446 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e68e246-bb9b-4b78-b668-ddcc38a6b26f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://497f797c94f36e14d2ed6ed0730a010d7c642a3148d1452b0a09da93269e4cab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://747254f43f4226e683b136aa920365af1759d867865aeb6625379211a983834d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T00:08:35Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 00:08:05.013265 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 00:08:05.015776 1 observer_polling.go:159] Starting file observer\\\\nI0318 00:08:05.046728 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 00:08:05.051577 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nF0318 00:08:35.404406 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:08:34Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93323833b9029d2087af6ebfab31298fc2405afdee386549b299ee95da60d26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76e713589e9d55c73767659fdc264925984ee7493cd62e37f4308a522e82d40b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:02 crc kubenswrapper[4983]: E0318 00:11:02.975040 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:11:02 crc kubenswrapper[4983]: I0318 00:11:02.984207 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5c5e541142d40b5a13956a771b0480b017b2a041df441fa1a7600a47c95722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:02Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.003286 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.032986 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.056361 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fd399e1a-84a7-410c-8977-5590886b1ecd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92c54fad4b50e832c30544eb1e07c034bc22fdeeb4a26887b92c4638bb01ef13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0562ba82e8bd277ed0e19df4b147b8cfc5121318388662d79812eca5b9a76d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef11882ecf742eaeeae92f73b2f02d55cb9a6712d43d703aa5d19b8a24e13fd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67719f21834b809979aba90cb1318fc89640c40e99cc0535ea94534dbad762d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e43f6c70f603d617e1d1d2fb1ba8d69249261549306d6cf9e8a84ec3a9810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af336f335c920f3d1c5ba21348759229a64816c8e86623db2814c8aebda00cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d2759bc17a5c2bfba402791540e28b9b17e67ffcc4bf52be10d4fb76c1c4163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dea1612578c883bfcf1ba7c3bfe6532f2fc89ac2e3c1196518069cf3bba6c5fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.071590 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"19860e89-0f4a-4c9f-b616-8329f0b9bab8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0befc5603e720538a3a0e9c40a50792f3e8d07c61929b87bd424427be0e54ddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655e9f142d07c3af06937170bb4eaa3ec5fbbb608685854cc1545665aec22b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://655e9f142d07c3af06937170bb4eaa3ec5fbbb608685854cc1545665aec22b24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.087965 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cvs9w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fe29e-1efb-4f9c-b0e4-ca34025d56de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb480465b084cf5a7a1e502271221f594a076dba66902fcd6226b546ef35797c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gw7t4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cvs9w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.108005 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6nblt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0853f8de-9bfc-4053-acd0-8893122d2d62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:30Z\\\",\\\"message\\\":\\\"2026-03-18T00:09:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c\\\\n2026-03-18T00:09:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863b9350-1807-4ea8-96b4-85aa2c434e5c to /host/opt/cni/bin/\\\\n2026-03-18T00:09:45Z [verbose] multus-daemon started\\\\n2026-03-18T00:09:45Z [verbose] Readiness Indicator file check\\\\n2026-03-18T00:10:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4vmw7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6nblt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.126165 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q9xgh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aaf037a1-6720-45d2-ada2-bb328bd16e04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cd36081749a010f9ad1cb9ed4dca360961a38e13aa574c3352fc121f15c6bd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wnmsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q9xgh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.146300 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fa65674-b13b-4d23-8fe9-9b677435bcd2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://064b97db9b50bd1e0de579a156b3ae410d90e954e3f6e0d19ce9123b12bfba6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbc10839ecb64952e617f23f96074d258d865a3dd53e34806c5a264fca597d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xh2c8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-67wt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.164735 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2cade86f-9f25-4f96-b861-8857ce2ad162\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:08:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81167a1f7781d69d700f7647928cbb027a40efec3129a860f62f82b12e2d9656\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d84f4ccd83826ef23c1583cae04ee8092f94bb5a2f9c948f97e503a9b1858201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3939424d6655bea60e9bad967f755a313411e8a699563cd3a9ef2b2e2354362\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd7a41aeff25b4185b6c4da1650da4a7807aa9aa3ca3d29bcfd23c147813d88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:08:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:08:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.184653 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d5498a411343debe30410243d0f84645a6593e74019a0ac4679699f2306f28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b02214c2b5d6b9ebc29db5686ced1cd46090f0898f05ce21d33bda53e805877a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.205825 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:27Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.229891 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"938043fe-450c-4477-8f55-312a843b0706\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T00:10:40Z\\\",\\\"message\\\":\\\"7 7553 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889001 7553 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889149 7553 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889194 7553 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 00:10:40.889258 7553 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 00:10:40.889459 7553 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0318 00:10:40.890665 7553 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T00:10:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T00:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T00:09:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T00:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj4g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4r8sk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.245572 4983 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe22dcea-cd03-43e3-80da-61a5f1e063ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T00:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mk9q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T00:09:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vjcg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T00:11:03Z is after 2025-08-24T17:21:41Z" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.843668 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.843761 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.843769 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:03 crc kubenswrapper[4983]: I0318 00:11:03.843857 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:03 crc kubenswrapper[4983]: E0318 00:11:03.844003 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:03 crc kubenswrapper[4983]: E0318 00:11:03.844202 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:03 crc kubenswrapper[4983]: E0318 00:11:03.844251 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:03 crc kubenswrapper[4983]: E0318 00:11:03.844329 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:05 crc kubenswrapper[4983]: I0318 00:11:05.843880 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:05 crc kubenswrapper[4983]: I0318 00:11:05.843978 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:05 crc kubenswrapper[4983]: E0318 00:11:05.844042 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:05 crc kubenswrapper[4983]: E0318 00:11:05.844212 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:05 crc kubenswrapper[4983]: I0318 00:11:05.844270 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:05 crc kubenswrapper[4983]: I0318 00:11:05.844282 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:05 crc kubenswrapper[4983]: E0318 00:11:05.844831 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:05 crc kubenswrapper[4983]: E0318 00:11:05.845035 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:05 crc kubenswrapper[4983]: I0318 00:11:05.845415 4983 scope.go:117] "RemoveContainer" containerID="2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e" Mar 18 00:11:05 crc kubenswrapper[4983]: E0318 00:11:05.845656 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" Mar 18 00:11:07 crc kubenswrapper[4983]: I0318 00:11:07.843663 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:07 crc kubenswrapper[4983]: I0318 00:11:07.843764 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:07 crc kubenswrapper[4983]: E0318 00:11:07.843796 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:07 crc kubenswrapper[4983]: I0318 00:11:07.843848 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:07 crc kubenswrapper[4983]: I0318 00:11:07.844111 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:07 crc kubenswrapper[4983]: E0318 00:11:07.844323 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:07 crc kubenswrapper[4983]: E0318 00:11:07.844504 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:07 crc kubenswrapper[4983]: E0318 00:11:07.844587 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:07 crc kubenswrapper[4983]: E0318 00:11:07.975876 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:11:09 crc kubenswrapper[4983]: I0318 00:11:09.844153 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:09 crc kubenswrapper[4983]: I0318 00:11:09.844244 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:09 crc kubenswrapper[4983]: E0318 00:11:09.844320 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:09 crc kubenswrapper[4983]: I0318 00:11:09.844349 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:09 crc kubenswrapper[4983]: I0318 00:11:09.844378 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:09 crc kubenswrapper[4983]: E0318 00:11:09.844566 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:09 crc kubenswrapper[4983]: E0318 00:11:09.844688 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:09 crc kubenswrapper[4983]: E0318 00:11:09.845043 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:11 crc kubenswrapper[4983]: I0318 00:11:11.844216 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:11 crc kubenswrapper[4983]: I0318 00:11:11.844232 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:11 crc kubenswrapper[4983]: I0318 00:11:11.844439 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:11 crc kubenswrapper[4983]: E0318 00:11:11.844568 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:11 crc kubenswrapper[4983]: E0318 00:11:11.844920 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:11 crc kubenswrapper[4983]: I0318 00:11:11.845137 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:11 crc kubenswrapper[4983]: E0318 00:11:11.845682 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:11 crc kubenswrapper[4983]: E0318 00:11:11.845969 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.321693 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.322263 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.322446 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.322634 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.322795 4983 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T00:11:12Z","lastTransitionTime":"2026-03-18T00:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.401255 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5"] Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.401810 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.405248 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.405452 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.405558 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.406839 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.428796 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-6nblt" podStartSLOduration=128.428761478 podStartE2EDuration="2m8.428761478s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:12.428060206 +0000 UTC m=+190.325434846" watchObservedRunningTime="2026-03-18 00:11:12.428761478 +0000 UTC m=+190.326136128" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.475397 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=94.475377103 podStartE2EDuration="1m34.475377103s" podCreationTimestamp="2026-03-18 00:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:12.473720931 +0000 UTC m=+190.371095571" watchObservedRunningTime="2026-03-18 00:11:12.475377103 +0000 UTC m=+190.372751713" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.475629 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-q9xgh" podStartSLOduration=128.475623911 podStartE2EDuration="2m8.475623911s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:12.445656869 +0000 UTC m=+190.343031499" watchObservedRunningTime="2026-03-18 00:11:12.475623911 +0000 UTC m=+190.372998521" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.499679 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=28.499656596 podStartE2EDuration="28.499656596s" podCreationTimestamp="2026-03-18 00:10:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:12.488705392 +0000 UTC m=+190.386080022" watchObservedRunningTime="2026-03-18 00:11:12.499656596 +0000 UTC m=+190.397031196" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.507256 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fa5676a-c793-411e-afff-40be97976373-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.507327 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7fa5676a-c793-411e-afff-40be97976373-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.507384 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7fa5676a-c793-411e-afff-40be97976373-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.507435 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7fa5676a-c793-411e-afff-40be97976373-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.507493 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7fa5676a-c793-411e-afff-40be97976373-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.518752 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-67wt2" podStartSLOduration=127.518730225 podStartE2EDuration="2m7.518730225s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:12.518583141 +0000 UTC m=+190.415957771" watchObservedRunningTime="2026-03-18 00:11:12.518730225 +0000 UTC m=+190.416104835" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.518889 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-cvs9w" podStartSLOduration=128.5188841 podStartE2EDuration="2m8.5188841s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:12.500488662 +0000 UTC m=+190.397863292" watchObservedRunningTime="2026-03-18 00:11:12.5188841 +0000 UTC m=+190.416258710" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.580382 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=57.580348602 podStartE2EDuration="57.580348602s" podCreationTimestamp="2026-03-18 00:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:12.579906468 +0000 UTC m=+190.477281118" watchObservedRunningTime="2026-03-18 00:11:12.580348602 +0000 UTC m=+190.477723222" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.608750 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7fa5676a-c793-411e-afff-40be97976373-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.609010 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7fa5676a-c793-411e-afff-40be97976373-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.609054 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fa5676a-c793-411e-afff-40be97976373-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.609156 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7fa5676a-c793-411e-afff-40be97976373-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.609279 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7fa5676a-c793-411e-afff-40be97976373-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.609351 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7fa5676a-c793-411e-afff-40be97976373-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.609396 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7fa5676a-c793-411e-afff-40be97976373-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.610051 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7fa5676a-c793-411e-afff-40be97976373-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.615094 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fa5676a-c793-411e-afff-40be97976373-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.632212 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7fa5676a-c793-411e-afff-40be97976373-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dz2d5\" (UID: \"7fa5676a-c793-411e-afff-40be97976373\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.714779 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podStartSLOduration=128.714757306 podStartE2EDuration="2m8.714757306s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:12.714640352 +0000 UTC m=+190.612014972" watchObservedRunningTime="2026-03-18 00:11:12.714757306 +0000 UTC m=+190.612131916" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.720555 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.747562 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-69n75" podStartSLOduration=128.747534006 podStartE2EDuration="2m8.747534006s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:12.745154121 +0000 UTC m=+190.642528741" watchObservedRunningTime="2026-03-18 00:11:12.747534006 +0000 UTC m=+190.644908626" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.781756 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=94.781736281 podStartE2EDuration="1m34.781736281s" podCreationTimestamp="2026-03-18 00:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:12.780831333 +0000 UTC m=+190.678205943" watchObservedRunningTime="2026-03-18 00:11:12.781736281 +0000 UTC m=+190.679110891" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.800413 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=64.800398158 podStartE2EDuration="1m4.800398158s" podCreationTimestamp="2026-03-18 00:10:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:12.799358675 +0000 UTC m=+190.696733315" watchObservedRunningTime="2026-03-18 00:11:12.800398158 +0000 UTC m=+190.697772768" Mar 18 00:11:12 crc kubenswrapper[4983]: E0318 00:11:12.977025 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:11:12 crc kubenswrapper[4983]: I0318 00:11:12.992332 4983 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 18 00:11:13 crc kubenswrapper[4983]: I0318 00:11:13.003766 4983 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 18 00:11:13 crc kubenswrapper[4983]: I0318 00:11:13.037177 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" event={"ID":"7fa5676a-c793-411e-afff-40be97976373","Type":"ContainerStarted","Data":"d58a6b9d6df958d129e9cfc042400093583c9fa7e500eb3ff06010a42d8c9e22"} Mar 18 00:11:13 crc kubenswrapper[4983]: I0318 00:11:13.037276 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" event={"ID":"7fa5676a-c793-411e-afff-40be97976373","Type":"ContainerStarted","Data":"94b481f6fe587c0580d94e0fac88c6e29d992374cae82b728dc6319a14ccd226"} Mar 18 00:11:13 crc kubenswrapper[4983]: I0318 00:11:13.843968 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:13 crc kubenswrapper[4983]: I0318 00:11:13.843979 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:13 crc kubenswrapper[4983]: E0318 00:11:13.844208 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:13 crc kubenswrapper[4983]: E0318 00:11:13.844369 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:13 crc kubenswrapper[4983]: I0318 00:11:13.844616 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:13 crc kubenswrapper[4983]: E0318 00:11:13.844760 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:13 crc kubenswrapper[4983]: I0318 00:11:13.844996 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:13 crc kubenswrapper[4983]: E0318 00:11:13.845435 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:15 crc kubenswrapper[4983]: I0318 00:11:15.844229 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:15 crc kubenswrapper[4983]: I0318 00:11:15.844341 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:15 crc kubenswrapper[4983]: E0318 00:11:15.845385 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:15 crc kubenswrapper[4983]: I0318 00:11:15.844441 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:15 crc kubenswrapper[4983]: E0318 00:11:15.845565 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:15 crc kubenswrapper[4983]: I0318 00:11:15.844381 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:15 crc kubenswrapper[4983]: E0318 00:11:15.845770 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:15 crc kubenswrapper[4983]: E0318 00:11:15.845860 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:17 crc kubenswrapper[4983]: I0318 00:11:17.053914 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6nblt_0853f8de-9bfc-4053-acd0-8893122d2d62/kube-multus/1.log" Mar 18 00:11:17 crc kubenswrapper[4983]: I0318 00:11:17.054541 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6nblt_0853f8de-9bfc-4053-acd0-8893122d2d62/kube-multus/0.log" Mar 18 00:11:17 crc kubenswrapper[4983]: I0318 00:11:17.054597 4983 generic.go:334] "Generic (PLEG): container finished" podID="0853f8de-9bfc-4053-acd0-8893122d2d62" containerID="96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b" exitCode=1 Mar 18 00:11:17 crc kubenswrapper[4983]: I0318 00:11:17.054630 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6nblt" event={"ID":"0853f8de-9bfc-4053-acd0-8893122d2d62","Type":"ContainerDied","Data":"96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b"} Mar 18 00:11:17 crc kubenswrapper[4983]: I0318 00:11:17.054669 4983 scope.go:117] "RemoveContainer" containerID="ed34f3efaa979fbd31d08ced653d865a6275c342228d7d800add68c2c7f5ffd2" Mar 18 00:11:17 crc kubenswrapper[4983]: I0318 00:11:17.055133 4983 scope.go:117] "RemoveContainer" containerID="96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b" Mar 18 00:11:17 crc kubenswrapper[4983]: E0318 00:11:17.055315 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-6nblt_openshift-multus(0853f8de-9bfc-4053-acd0-8893122d2d62)\"" pod="openshift-multus/multus-6nblt" podUID="0853f8de-9bfc-4053-acd0-8893122d2d62" Mar 18 00:11:17 crc kubenswrapper[4983]: I0318 00:11:17.077837 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz2d5" podStartSLOduration=133.077816687 podStartE2EDuration="2m13.077816687s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:13.061573285 +0000 UTC m=+190.958947915" watchObservedRunningTime="2026-03-18 00:11:17.077816687 +0000 UTC m=+194.975191307" Mar 18 00:11:17 crc kubenswrapper[4983]: I0318 00:11:17.843438 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:17 crc kubenswrapper[4983]: I0318 00:11:17.843476 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:17 crc kubenswrapper[4983]: I0318 00:11:17.843430 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:17 crc kubenswrapper[4983]: E0318 00:11:17.843610 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:17 crc kubenswrapper[4983]: I0318 00:11:17.843621 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:17 crc kubenswrapper[4983]: E0318 00:11:17.843785 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:17 crc kubenswrapper[4983]: E0318 00:11:17.843994 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:17 crc kubenswrapper[4983]: E0318 00:11:17.844079 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:17 crc kubenswrapper[4983]: I0318 00:11:17.845296 4983 scope.go:117] "RemoveContainer" containerID="2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e" Mar 18 00:11:17 crc kubenswrapper[4983]: E0318 00:11:17.845635 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4r8sk_openshift-ovn-kubernetes(938043fe-450c-4477-8f55-312a843b0706)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" Mar 18 00:11:17 crc kubenswrapper[4983]: E0318 00:11:17.979539 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:11:18 crc kubenswrapper[4983]: I0318 00:11:18.093703 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6nblt_0853f8de-9bfc-4053-acd0-8893122d2d62/kube-multus/1.log" Mar 18 00:11:19 crc kubenswrapper[4983]: I0318 00:11:19.843504 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:19 crc kubenswrapper[4983]: I0318 00:11:19.843574 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:19 crc kubenswrapper[4983]: E0318 00:11:19.844286 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:19 crc kubenswrapper[4983]: I0318 00:11:19.843709 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:19 crc kubenswrapper[4983]: I0318 00:11:19.843607 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:19 crc kubenswrapper[4983]: E0318 00:11:19.844445 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:19 crc kubenswrapper[4983]: E0318 00:11:19.844687 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:19 crc kubenswrapper[4983]: E0318 00:11:19.844824 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:21 crc kubenswrapper[4983]: I0318 00:11:21.843625 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:21 crc kubenswrapper[4983]: I0318 00:11:21.843672 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:21 crc kubenswrapper[4983]: I0318 00:11:21.843659 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:21 crc kubenswrapper[4983]: E0318 00:11:21.843850 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:21 crc kubenswrapper[4983]: I0318 00:11:21.843866 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:21 crc kubenswrapper[4983]: E0318 00:11:21.844002 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:21 crc kubenswrapper[4983]: E0318 00:11:21.844109 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:21 crc kubenswrapper[4983]: E0318 00:11:21.844165 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:22 crc kubenswrapper[4983]: E0318 00:11:22.984065 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:11:23 crc kubenswrapper[4983]: I0318 00:11:23.843649 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:23 crc kubenswrapper[4983]: I0318 00:11:23.843699 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:23 crc kubenswrapper[4983]: I0318 00:11:23.843783 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:23 crc kubenswrapper[4983]: I0318 00:11:23.843667 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:23 crc kubenswrapper[4983]: E0318 00:11:23.843854 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:23 crc kubenswrapper[4983]: E0318 00:11:23.844072 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:23 crc kubenswrapper[4983]: E0318 00:11:23.844299 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:23 crc kubenswrapper[4983]: E0318 00:11:23.844371 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:25 crc kubenswrapper[4983]: I0318 00:11:25.843784 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:25 crc kubenswrapper[4983]: I0318 00:11:25.843982 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:25 crc kubenswrapper[4983]: I0318 00:11:25.844013 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:25 crc kubenswrapper[4983]: I0318 00:11:25.844143 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:25 crc kubenswrapper[4983]: E0318 00:11:25.844258 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:25 crc kubenswrapper[4983]: E0318 00:11:25.844549 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:25 crc kubenswrapper[4983]: E0318 00:11:25.844645 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:25 crc kubenswrapper[4983]: E0318 00:11:25.844959 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:27 crc kubenswrapper[4983]: I0318 00:11:27.843795 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:27 crc kubenswrapper[4983]: I0318 00:11:27.843861 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:27 crc kubenswrapper[4983]: I0318 00:11:27.843836 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:27 crc kubenswrapper[4983]: I0318 00:11:27.843981 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:27 crc kubenswrapper[4983]: E0318 00:11:27.844033 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:27 crc kubenswrapper[4983]: E0318 00:11:27.844287 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:27 crc kubenswrapper[4983]: I0318 00:11:27.844735 4983 scope.go:117] "RemoveContainer" containerID="96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b" Mar 18 00:11:27 crc kubenswrapper[4983]: E0318 00:11:27.844773 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:27 crc kubenswrapper[4983]: E0318 00:11:27.844971 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:27 crc kubenswrapper[4983]: E0318 00:11:27.985751 4983 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 00:11:28 crc kubenswrapper[4983]: I0318 00:11:28.133806 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6nblt_0853f8de-9bfc-4053-acd0-8893122d2d62/kube-multus/1.log" Mar 18 00:11:28 crc kubenswrapper[4983]: I0318 00:11:28.133861 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6nblt" event={"ID":"0853f8de-9bfc-4053-acd0-8893122d2d62","Type":"ContainerStarted","Data":"106e6e771cbb1f79ee805368b6caaefd9fb0ec1fa69685d104bc832621c8f39d"} Mar 18 00:11:28 crc kubenswrapper[4983]: I0318 00:11:28.845617 4983 scope.go:117] "RemoveContainer" containerID="2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e" Mar 18 00:11:29 crc kubenswrapper[4983]: I0318 00:11:29.140823 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/3.log" Mar 18 00:11:29 crc kubenswrapper[4983]: I0318 00:11:29.144556 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerStarted","Data":"3f11bbbdc7608e518cf73e0bbbff4326a4268dec9707d318c5af979798250551"} Mar 18 00:11:29 crc kubenswrapper[4983]: I0318 00:11:29.145156 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:11:29 crc kubenswrapper[4983]: I0318 00:11:29.178511 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podStartSLOduration=144.178490624 podStartE2EDuration="2m24.178490624s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:29.177449833 +0000 UTC m=+207.074824463" watchObservedRunningTime="2026-03-18 00:11:29.178490624 +0000 UTC m=+207.075865244" Mar 18 00:11:29 crc kubenswrapper[4983]: I0318 00:11:29.702208 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vjcg5"] Mar 18 00:11:29 crc kubenswrapper[4983]: I0318 00:11:29.702430 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:29 crc kubenswrapper[4983]: E0318 00:11:29.702632 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:29 crc kubenswrapper[4983]: I0318 00:11:29.844461 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:29 crc kubenswrapper[4983]: I0318 00:11:29.844533 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:29 crc kubenswrapper[4983]: I0318 00:11:29.844547 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:29 crc kubenswrapper[4983]: E0318 00:11:29.844845 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:29 crc kubenswrapper[4983]: E0318 00:11:29.844888 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:29 crc kubenswrapper[4983]: E0318 00:11:29.845059 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:30 crc kubenswrapper[4983]: I0318 00:11:30.844522 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:30 crc kubenswrapper[4983]: E0318 00:11:30.844726 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:31 crc kubenswrapper[4983]: I0318 00:11:31.844363 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:31 crc kubenswrapper[4983]: I0318 00:11:31.844423 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:31 crc kubenswrapper[4983]: I0318 00:11:31.844384 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:31 crc kubenswrapper[4983]: E0318 00:11:31.844586 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 00:11:31 crc kubenswrapper[4983]: E0318 00:11:31.844663 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 00:11:31 crc kubenswrapper[4983]: E0318 00:11:31.844798 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 00:11:32 crc kubenswrapper[4983]: I0318 00:11:32.844198 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:32 crc kubenswrapper[4983]: E0318 00:11:32.844496 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vjcg5" podUID="fe22dcea-cd03-43e3-80da-61a5f1e063ad" Mar 18 00:11:33 crc kubenswrapper[4983]: I0318 00:11:33.844408 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:33 crc kubenswrapper[4983]: I0318 00:11:33.844492 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:33 crc kubenswrapper[4983]: I0318 00:11:33.844419 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:33 crc kubenswrapper[4983]: I0318 00:11:33.847325 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 18 00:11:33 crc kubenswrapper[4983]: I0318 00:11:33.847825 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 18 00:11:33 crc kubenswrapper[4983]: I0318 00:11:33.848101 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 18 00:11:33 crc kubenswrapper[4983]: I0318 00:11:33.848154 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 18 00:11:34 crc kubenswrapper[4983]: I0318 00:11:34.844212 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:11:34 crc kubenswrapper[4983]: I0318 00:11:34.847081 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 18 00:11:34 crc kubenswrapper[4983]: I0318 00:11:34.847363 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 18 00:11:35 crc kubenswrapper[4983]: I0318 00:11:35.837483 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:35 crc kubenswrapper[4983]: I0318 00:11:35.837632 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:35 crc kubenswrapper[4983]: I0318 00:11:35.837690 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:35 crc kubenswrapper[4983]: I0318 00:11:35.837760 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:35 crc kubenswrapper[4983]: I0318 00:11:35.837778 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:35 crc kubenswrapper[4983]: E0318 00:11:35.837831 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:13:37.837786671 +0000 UTC m=+335.735161311 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:35 crc kubenswrapper[4983]: I0318 00:11:35.839185 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:35 crc kubenswrapper[4983]: I0318 00:11:35.845190 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:35 crc kubenswrapper[4983]: I0318 00:11:35.845522 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:35 crc kubenswrapper[4983]: I0318 00:11:35.845752 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:35 crc kubenswrapper[4983]: I0318 00:11:35.971876 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 00:11:35 crc kubenswrapper[4983]: I0318 00:11:35.991195 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:36 crc kubenswrapper[4983]: I0318 00:11:36.004621 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 00:11:36 crc kubenswrapper[4983]: W0318 00:11:36.199502 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-f4c8ab5b7e780d08406c9aa31374b5b4ce540b13742b2e6a6551eea8aa59fc6e WatchSource:0}: Error finding container f4c8ab5b7e780d08406c9aa31374b5b4ce540b13742b2e6a6551eea8aa59fc6e: Status 404 returned error can't find the container with id f4c8ab5b7e780d08406c9aa31374b5b4ce540b13742b2e6a6551eea8aa59fc6e Mar 18 00:11:36 crc kubenswrapper[4983]: W0318 00:11:36.256538 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-0ea852ebe6e7f74c91ac926204882de442969ec23bbdffe8b5259853889cea8c WatchSource:0}: Error finding container 0ea852ebe6e7f74c91ac926204882de442969ec23bbdffe8b5259853889cea8c: Status 404 returned error can't find the container with id 0ea852ebe6e7f74c91ac926204882de442969ec23bbdffe8b5259853889cea8c Mar 18 00:11:37 crc kubenswrapper[4983]: I0318 00:11:37.174222 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"67d07cee3c66230fabf17ec9ad8e3a1987a492c55120165245635a5d594116a3"} Mar 18 00:11:37 crc kubenswrapper[4983]: I0318 00:11:37.174334 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0ea852ebe6e7f74c91ac926204882de442969ec23bbdffe8b5259853889cea8c"} Mar 18 00:11:37 crc kubenswrapper[4983]: I0318 00:11:37.174542 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:11:37 crc kubenswrapper[4983]: I0318 00:11:37.177039 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"309c0c922ee722fec745ab03189e9785781c5659096098d1fe1328f1779294c5"} Mar 18 00:11:37 crc kubenswrapper[4983]: I0318 00:11:37.177083 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0a9df467bf45484f68e12fa47fdd76e29737fdc6904aa399b3d861230e08fe8c"} Mar 18 00:11:37 crc kubenswrapper[4983]: I0318 00:11:37.179158 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"663ba8243ab667a7b6413596cb3305ad067e8c0fa499f6a1cae24aab8a5d70f5"} Mar 18 00:11:37 crc kubenswrapper[4983]: I0318 00:11:37.179202 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f4c8ab5b7e780d08406c9aa31374b5b4ce540b13742b2e6a6551eea8aa59fc6e"} Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.043211 4983 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.106376 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.107354 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.108047 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4dn4m"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.108896 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.110851 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.111470 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h74lp"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.111631 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.112541 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.113297 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.113526 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.113675 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.114105 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.114529 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.115034 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.115660 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.115879 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.116071 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.116443 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.116532 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zlqph"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.117184 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.117218 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.120247 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.120641 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.123844 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.124798 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.130408 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.130683 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.130890 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.131117 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.131212 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.131231 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.131342 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.131359 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.131631 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.131964 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.132150 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.132312 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.132632 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.132777 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.145960 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.147532 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.148470 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.149338 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.149639 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.150710 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.151751 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.166267 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.167222 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.168058 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-w4s69"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.169024 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-w4s69" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.169460 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-slqc5"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.172209 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.172376 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.172496 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.172663 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.172803 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.172823 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.173028 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.173141 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.173709 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.173817 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.174014 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.174686 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.175157 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.175460 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.176010 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.179151 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.179405 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-z7h4m"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.180227 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.182920 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.183204 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.184147 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.184758 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29563200-6l8ds"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.185064 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29563200-6l8ds" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.185210 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.185520 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.186246 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.186350 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.187028 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-gvd97"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.188316 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.192715 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.193325 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.193635 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.193803 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-65l25"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.193883 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.194175 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.194326 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.194431 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.194627 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.194748 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.194848 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.194949 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.196885 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6t6lv"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.197099 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.197369 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4dn4m"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.197466 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.205363 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.205602 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.205798 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.206811 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.207011 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.207148 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.207400 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.207560 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.207694 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.207891 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.210177 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.210490 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.210672 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.211399 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.211616 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.211758 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.211920 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.212046 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.212257 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.213391 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.215373 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.219024 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.238714 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66e8121d-9422-4b23-93e9-9f4d350c149f-audit-dir\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.238757 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.238777 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/651bef83-00db-40d1-b096-247b3c645859-config\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.238800 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c447db7-c45e-4e01-925a-6ac56ec2aee2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-l46js\" (UID: \"5c447db7-c45e-4e01-925a-6ac56ec2aee2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.238829 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tpgg\" (UniqueName: \"kubernetes.io/projected/0774a9cb-d9cb-45a2-89e1-384902cf8f94-kube-api-access-4tpgg\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.238848 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-etcd-serving-ca\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.238870 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.238894 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2p82\" (UniqueName: \"kubernetes.io/projected/66e8121d-9422-4b23-93e9-9f4d350c149f-kube-api-access-m2p82\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.239009 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxm5m\" (UniqueName: \"kubernetes.io/projected/4fbfc372-868f-41d5-81d0-efb8a3e88023-kube-api-access-dxm5m\") pod \"machine-api-operator-5694c8668f-4dn4m\" (UID: \"4fbfc372-868f-41d5-81d0-efb8a3e88023\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.239217 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/53bc5818-4983-4f5f-8a3c-edc02aff5325-encryption-config\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.239401 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0774a9cb-d9cb-45a2-89e1-384902cf8f94-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.239610 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79f09f2f-fe8b-4340-a1fc-c16f212b7e31-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2xj85\" (UID: \"79f09f2f-fe8b-4340-a1fc-c16f212b7e31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.239788 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.239735 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/651bef83-00db-40d1-b096-247b3c645859-etcd-service-ca\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.239869 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4887x\" (UniqueName: \"kubernetes.io/projected/5c447db7-c45e-4e01-925a-6ac56ec2aee2-kube-api-access-4887x\") pod \"openshift-apiserver-operator-796bbdcf4f-l46js\" (UID: \"5c447db7-c45e-4e01-925a-6ac56ec2aee2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.240009 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0774a9cb-d9cb-45a2-89e1-384902cf8f94-serving-cert\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.240104 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79f09f2f-fe8b-4340-a1fc-c16f212b7e31-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2xj85\" (UID: \"79f09f2f-fe8b-4340-a1fc-c16f212b7e31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.241065 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53bc5818-4983-4f5f-8a3c-edc02aff5325-serving-cert\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.241103 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.241130 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.241201 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6e69d708-f2df-436e-a5ec-41848aecbace-metrics-tls\") pod \"dns-operator-744455d44c-w4s69\" (UID: \"6e69d708-f2df-436e-a5ec-41848aecbace\") " pod="openshift-dns-operator/dns-operator-744455d44c-w4s69" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.241315 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0774a9cb-d9cb-45a2-89e1-384902cf8f94-audit-policies\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.241445 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04baa777-3a33-4b5a-9808-99a84b1b8005-config\") pod \"console-operator-58897d9998-slqc5\" (UID: \"04baa777-3a33-4b5a-9808-99a84b1b8005\") " pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.241615 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/04baa777-3a33-4b5a-9808-99a84b1b8005-trusted-ca\") pod \"console-operator-58897d9998-slqc5\" (UID: \"04baa777-3a33-4b5a-9808-99a84b1b8005\") " pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.242828 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-config\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.243120 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-config\") pod \"route-controller-manager-6576b87f9c-5j7r4\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.243227 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-config\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.243314 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04baa777-3a33-4b5a-9808-99a84b1b8005-serving-cert\") pod \"console-operator-58897d9998-slqc5\" (UID: \"04baa777-3a33-4b5a-9808-99a84b1b8005\") " pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.243341 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-service-ca-bundle\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.243504 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq7j2\" (UniqueName: \"kubernetes.io/projected/0b2013fc-d64f-4f21-be62-e17f3855f6c1-kube-api-access-lq7j2\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.243534 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0774a9cb-d9cb-45a2-89e1-384902cf8f94-encryption-config\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.243598 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqcc9\" (UniqueName: \"kubernetes.io/projected/6e69d708-f2df-436e-a5ec-41848aecbace-kube-api-access-zqcc9\") pod \"dns-operator-744455d44c-w4s69\" (UID: \"6e69d708-f2df-436e-a5ec-41848aecbace\") " pod="openshift-dns-operator/dns-operator-744455d44c-w4s69" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.243712 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53bc5818-4983-4f5f-8a3c-edc02aff5325-audit-dir\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.243845 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/744c22bb-2f35-494b-b352-ff71691bd0af-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pg4wt\" (UID: \"744c22bb-2f35-494b-b352-ff71691bd0af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.243906 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244038 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6375157a-012b-441e-b97b-39d6830cef68-console-config\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244118 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-serving-cert\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244152 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/651bef83-00db-40d1-b096-247b3c645859-etcd-ca\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244189 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/744c22bb-2f35-494b-b352-ff71691bd0af-serving-cert\") pod \"openshift-config-operator-7777fb866f-pg4wt\" (UID: \"744c22bb-2f35-494b-b352-ff71691bd0af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244293 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4fbfc372-868f-41d5-81d0-efb8a3e88023-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4dn4m\" (UID: \"4fbfc372-868f-41d5-81d0-efb8a3e88023\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244485 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/812c885e-98f8-47cb-afd6-5fb537bea3e3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-w8wsj\" (UID: \"812c885e-98f8-47cb-afd6-5fb537bea3e3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244587 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqwc6\" (UniqueName: \"kubernetes.io/projected/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-kube-api-access-vqwc6\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244659 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-audit\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244714 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6375157a-012b-441e-b97b-39d6830cef68-console-serving-cert\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244739 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0774a9cb-d9cb-45a2-89e1-384902cf8f94-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244755 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0774a9cb-d9cb-45a2-89e1-384902cf8f94-audit-dir\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244779 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244835 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-client-ca\") pod \"route-controller-manager-6576b87f9c-5j7r4\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244857 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf67f\" (UniqueName: \"kubernetes.io/projected/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-kube-api-access-nf67f\") pod \"route-controller-manager-6576b87f9c-5j7r4\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244877 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/53bc5818-4983-4f5f-8a3c-edc02aff5325-etcd-client\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.244897 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c447db7-c45e-4e01-925a-6ac56ec2aee2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-l46js\" (UID: \"5c447db7-c45e-4e01-925a-6ac56ec2aee2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.267472 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.268020 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.268077 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269210 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b2013fc-d64f-4f21-be62-e17f3855f6c1-serving-cert\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269272 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/53bc5818-4983-4f5f-8a3c-edc02aff5325-node-pullsecrets\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269296 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-audit-policies\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269317 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6375157a-012b-441e-b97b-39d6830cef68-trusted-ca-bundle\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269340 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4fbfc372-868f-41d5-81d0-efb8a3e88023-images\") pod \"machine-api-operator-5694c8668f-4dn4m\" (UID: \"4fbfc372-868f-41d5-81d0-efb8a3e88023\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269362 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/651bef83-00db-40d1-b096-247b3c645859-etcd-client\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269382 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqntz\" (UniqueName: \"kubernetes.io/projected/651bef83-00db-40d1-b096-247b3c645859-kube-api-access-wqntz\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269401 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8-serviceca\") pod \"image-pruner-29563200-6l8ds\" (UID: \"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8\") " pod="openshift-image-registry/image-pruner-29563200-6l8ds" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269424 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6375157a-012b-441e-b97b-39d6830cef68-console-oauth-config\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269444 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6375157a-012b-441e-b97b-39d6830cef68-oauth-serving-cert\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269465 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-config\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269485 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-client-ca\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269505 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pkkq\" (UniqueName: \"kubernetes.io/projected/744c22bb-2f35-494b-b352-ff71691bd0af-kube-api-access-6pkkq\") pod \"openshift-config-operator-7777fb866f-pg4wt\" (UID: \"744c22bb-2f35-494b-b352-ff71691bd0af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269528 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269552 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269573 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tgpb\" (UniqueName: \"kubernetes.io/projected/812c885e-98f8-47cb-afd6-5fb537bea3e3-kube-api-access-8tgpb\") pod \"cluster-samples-operator-665b6dd947-w8wsj\" (UID: \"812c885e-98f8-47cb-afd6-5fb537bea3e3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269595 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269619 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269661 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kcsp\" (UniqueName: \"kubernetes.io/projected/6375157a-012b-441e-b97b-39d6830cef68-kube-api-access-5kcsp\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269681 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269702 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx7hh\" (UniqueName: \"kubernetes.io/projected/79f09f2f-fe8b-4340-a1fc-c16f212b7e31-kube-api-access-gx7hh\") pod \"openshift-controller-manager-operator-756b6f6bc6-2xj85\" (UID: \"79f09f2f-fe8b-4340-a1fc-c16f212b7e31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269723 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-serving-cert\") pod \"route-controller-manager-6576b87f9c-5j7r4\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269745 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2ms6\" (UniqueName: \"kubernetes.io/projected/c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8-kube-api-access-t2ms6\") pod \"image-pruner-29563200-6l8ds\" (UID: \"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8\") " pod="openshift-image-registry/image-pruner-29563200-6l8ds" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269767 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0774a9cb-d9cb-45a2-89e1-384902cf8f94-etcd-client\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269787 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-image-import-ca\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269808 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-246g7\" (UniqueName: \"kubernetes.io/projected/53bc5818-4983-4f5f-8a3c-edc02aff5325-kube-api-access-246g7\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269828 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269854 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/651bef83-00db-40d1-b096-247b3c645859-serving-cert\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269875 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvvzr\" (UniqueName: \"kubernetes.io/projected/04baa777-3a33-4b5a-9808-99a84b1b8005-kube-api-access-lvvzr\") pod \"console-operator-58897d9998-slqc5\" (UID: \"04baa777-3a33-4b5a-9808-99a84b1b8005\") " pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269896 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6375157a-012b-441e-b97b-39d6830cef68-service-ca\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.269917 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fbfc372-868f-41d5-81d0-efb8a3e88023-config\") pod \"machine-api-operator-5694c8668f-4dn4m\" (UID: \"4fbfc372-868f-41d5-81d0-efb8a3e88023\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.270228 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.270408 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.270423 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.270478 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.270573 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.270591 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.270658 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.270838 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.271027 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.271133 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.271183 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.271282 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.271148 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.271530 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-44428"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.271683 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.272021 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.272092 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.272361 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.272694 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-87bh5"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.272894 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.273153 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zlqph"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.273317 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-87bh5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.275969 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.281756 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.282100 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.285816 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.289012 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.289731 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.290015 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.291499 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.291625 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-wxstd"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.292118 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wxstd" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.292384 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qkkt2"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.296732 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.298570 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.299030 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.299752 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rpftn"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.299966 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.300821 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.301140 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.302358 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.302493 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.302883 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-ncss9"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.303230 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.303456 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.303804 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.304414 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.304902 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-85cfx"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.305535 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-85cfx" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.307532 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-w4s69"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.311650 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.312374 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.312573 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.312735 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.313233 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.313824 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-f8svz"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.314418 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.319816 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gx8mh"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.320421 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.323095 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.323673 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.323883 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.325913 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563210-lptd5"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.328330 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563210-lptd5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.328637 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.330599 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.333382 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.334628 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.336390 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.336898 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.338134 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.338485 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.339193 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.341630 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-b8gtz"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.342291 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-pch42"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.342508 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b8gtz" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.346246 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.346279 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.346626 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-pch42" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.346984 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.347215 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-slqc5"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.348522 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29563200-6l8ds"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.349011 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.349813 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h74lp"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.351916 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.353498 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-gvd97"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.354951 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-65l25"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.356638 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-87bh5"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.359804 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.361340 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-k424z"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.362051 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-k424z" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.363237 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-kzks7"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.364233 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.364894 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.366478 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6t6lv"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.368201 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-z7h4m"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.370024 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.370268 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.371009 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-service-ca-bundle\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.371041 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq7j2\" (UniqueName: \"kubernetes.io/projected/0b2013fc-d64f-4f21-be62-e17f3855f6c1-kube-api-access-lq7j2\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.371070 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0774a9cb-d9cb-45a2-89e1-384902cf8f94-encryption-config\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.371093 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqcc9\" (UniqueName: \"kubernetes.io/projected/6e69d708-f2df-436e-a5ec-41848aecbace-kube-api-access-zqcc9\") pod \"dns-operator-744455d44c-w4s69\" (UID: \"6e69d708-f2df-436e-a5ec-41848aecbace\") " pod="openshift-dns-operator/dns-operator-744455d44c-w4s69" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.371115 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53bc5818-4983-4f5f-8a3c-edc02aff5325-audit-dir\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.371145 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/744c22bb-2f35-494b-b352-ff71691bd0af-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pg4wt\" (UID: \"744c22bb-2f35-494b-b352-ff71691bd0af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.371169 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.371192 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6375157a-012b-441e-b97b-39d6830cef68-console-config\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.371215 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-serving-cert\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.371238 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/651bef83-00db-40d1-b096-247b3c645859-etcd-ca\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.371306 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53bc5818-4983-4f5f-8a3c-edc02aff5325-audit-dir\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.371419 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.372034 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/744c22bb-2f35-494b-b352-ff71691bd0af-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pg4wt\" (UID: \"744c22bb-2f35-494b-b352-ff71691bd0af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.372687 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6375157a-012b-441e-b97b-39d6830cef68-console-config\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.372728 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/651bef83-00db-40d1-b096-247b3c645859-etcd-ca\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.372796 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373001 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/744c22bb-2f35-494b-b352-ff71691bd0af-serving-cert\") pod \"openshift-config-operator-7777fb866f-pg4wt\" (UID: \"744c22bb-2f35-494b-b352-ff71691bd0af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373039 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4fbfc372-868f-41d5-81d0-efb8a3e88023-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4dn4m\" (UID: \"4fbfc372-868f-41d5-81d0-efb8a3e88023\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373068 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/812c885e-98f8-47cb-afd6-5fb537bea3e3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-w8wsj\" (UID: \"812c885e-98f8-47cb-afd6-5fb537bea3e3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373095 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqwc6\" (UniqueName: \"kubernetes.io/projected/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-kube-api-access-vqwc6\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373117 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-audit\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373171 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6375157a-012b-441e-b97b-39d6830cef68-console-serving-cert\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373196 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0774a9cb-d9cb-45a2-89e1-384902cf8f94-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373217 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0774a9cb-d9cb-45a2-89e1-384902cf8f94-audit-dir\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373238 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373264 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-client-ca\") pod \"route-controller-manager-6576b87f9c-5j7r4\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373287 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf67f\" (UniqueName: \"kubernetes.io/projected/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-kube-api-access-nf67f\") pod \"route-controller-manager-6576b87f9c-5j7r4\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373309 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/53bc5818-4983-4f5f-8a3c-edc02aff5325-etcd-client\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373331 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c447db7-c45e-4e01-925a-6ac56ec2aee2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-l46js\" (UID: \"5c447db7-c45e-4e01-925a-6ac56ec2aee2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373885 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0774a9cb-d9cb-45a2-89e1-384902cf8f94-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.373040 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-service-ca-bundle\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374270 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6375157a-012b-441e-b97b-39d6830cef68-trusted-ca-bundle\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374306 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b2013fc-d64f-4f21-be62-e17f3855f6c1-serving-cert\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374348 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/53bc5818-4983-4f5f-8a3c-edc02aff5325-node-pullsecrets\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374369 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-audit-policies\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374391 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4fbfc372-868f-41d5-81d0-efb8a3e88023-images\") pod \"machine-api-operator-5694c8668f-4dn4m\" (UID: \"4fbfc372-868f-41d5-81d0-efb8a3e88023\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374421 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/651bef83-00db-40d1-b096-247b3c645859-etcd-client\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374447 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqntz\" (UniqueName: \"kubernetes.io/projected/651bef83-00db-40d1-b096-247b3c645859-kube-api-access-wqntz\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374468 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8-serviceca\") pod \"image-pruner-29563200-6l8ds\" (UID: \"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8\") " pod="openshift-image-registry/image-pruner-29563200-6l8ds" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374509 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6375157a-012b-441e-b97b-39d6830cef68-console-oauth-config\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374533 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6375157a-012b-441e-b97b-39d6830cef68-oauth-serving-cert\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374589 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-config\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374602 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-wxstd"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374612 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-client-ca\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374633 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pkkq\" (UniqueName: \"kubernetes.io/projected/744c22bb-2f35-494b-b352-ff71691bd0af-kube-api-access-6pkkq\") pod \"openshift-config-operator-7777fb866f-pg4wt\" (UID: \"744c22bb-2f35-494b-b352-ff71691bd0af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374659 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374683 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374708 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tgpb\" (UniqueName: \"kubernetes.io/projected/812c885e-98f8-47cb-afd6-5fb537bea3e3-kube-api-access-8tgpb\") pod \"cluster-samples-operator-665b6dd947-w8wsj\" (UID: \"812c885e-98f8-47cb-afd6-5fb537bea3e3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374734 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374760 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374797 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kcsp\" (UniqueName: \"kubernetes.io/projected/6375157a-012b-441e-b97b-39d6830cef68-kube-api-access-5kcsp\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374821 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374846 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx7hh\" (UniqueName: \"kubernetes.io/projected/79f09f2f-fe8b-4340-a1fc-c16f212b7e31-kube-api-access-gx7hh\") pod \"openshift-controller-manager-operator-756b6f6bc6-2xj85\" (UID: \"79f09f2f-fe8b-4340-a1fc-c16f212b7e31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374872 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-serving-cert\") pod \"route-controller-manager-6576b87f9c-5j7r4\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374911 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-client-ca\") pod \"route-controller-manager-6576b87f9c-5j7r4\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.374917 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2ms6\" (UniqueName: \"kubernetes.io/projected/c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8-kube-api-access-t2ms6\") pod \"image-pruner-29563200-6l8ds\" (UID: \"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8\") " pod="openshift-image-registry/image-pruner-29563200-6l8ds" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375006 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8qn5\" (UniqueName: \"kubernetes.io/projected/b876a42a-856b-4530-a601-3ca35db1078a-kube-api-access-v8qn5\") pod \"migrator-59844c95c7-wxstd\" (UID: \"b876a42a-856b-4530-a601-3ca35db1078a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wxstd" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375045 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0774a9cb-d9cb-45a2-89e1-384902cf8f94-etcd-client\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375069 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-image-import-ca\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375093 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-246g7\" (UniqueName: \"kubernetes.io/projected/53bc5818-4983-4f5f-8a3c-edc02aff5325-kube-api-access-246g7\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375118 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375149 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/651bef83-00db-40d1-b096-247b3c645859-serving-cert\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375171 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvvzr\" (UniqueName: \"kubernetes.io/projected/04baa777-3a33-4b5a-9808-99a84b1b8005-kube-api-access-lvvzr\") pod \"console-operator-58897d9998-slqc5\" (UID: \"04baa777-3a33-4b5a-9808-99a84b1b8005\") " pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375194 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6375157a-012b-441e-b97b-39d6830cef68-service-ca\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375219 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fbfc372-868f-41d5-81d0-efb8a3e88023-config\") pod \"machine-api-operator-5694c8668f-4dn4m\" (UID: \"4fbfc372-868f-41d5-81d0-efb8a3e88023\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375240 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375267 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375292 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375318 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66e8121d-9422-4b23-93e9-9f4d350c149f-audit-dir\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375344 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375369 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/651bef83-00db-40d1-b096-247b3c645859-config\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375393 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c447db7-c45e-4e01-925a-6ac56ec2aee2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-l46js\" (UID: \"5c447db7-c45e-4e01-925a-6ac56ec2aee2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375399 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4fbfc372-868f-41d5-81d0-efb8a3e88023-images\") pod \"machine-api-operator-5694c8668f-4dn4m\" (UID: \"4fbfc372-868f-41d5-81d0-efb8a3e88023\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375422 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tpgg\" (UniqueName: \"kubernetes.io/projected/0774a9cb-d9cb-45a2-89e1-384902cf8f94-kube-api-access-4tpgg\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375443 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-etcd-serving-ca\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375464 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375486 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2p82\" (UniqueName: \"kubernetes.io/projected/66e8121d-9422-4b23-93e9-9f4d350c149f-kube-api-access-m2p82\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375509 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxm5m\" (UniqueName: \"kubernetes.io/projected/4fbfc372-868f-41d5-81d0-efb8a3e88023-kube-api-access-dxm5m\") pod \"machine-api-operator-5694c8668f-4dn4m\" (UID: \"4fbfc372-868f-41d5-81d0-efb8a3e88023\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375535 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/53bc5818-4983-4f5f-8a3c-edc02aff5325-encryption-config\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375556 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0774a9cb-d9cb-45a2-89e1-384902cf8f94-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375580 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79f09f2f-fe8b-4340-a1fc-c16f212b7e31-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2xj85\" (UID: \"79f09f2f-fe8b-4340-a1fc-c16f212b7e31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375603 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/651bef83-00db-40d1-b096-247b3c645859-etcd-service-ca\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375607 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375624 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4887x\" (UniqueName: \"kubernetes.io/projected/5c447db7-c45e-4e01-925a-6ac56ec2aee2-kube-api-access-4887x\") pod \"openshift-apiserver-operator-796bbdcf4f-l46js\" (UID: \"5c447db7-c45e-4e01-925a-6ac56ec2aee2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375653 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0774a9cb-d9cb-45a2-89e1-384902cf8f94-serving-cert\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375672 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79f09f2f-fe8b-4340-a1fc-c16f212b7e31-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2xj85\" (UID: \"79f09f2f-fe8b-4340-a1fc-c16f212b7e31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375692 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6e69d708-f2df-436e-a5ec-41848aecbace-metrics-tls\") pod \"dns-operator-744455d44c-w4s69\" (UID: \"6e69d708-f2df-436e-a5ec-41848aecbace\") " pod="openshift-dns-operator/dns-operator-744455d44c-w4s69" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375716 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53bc5818-4983-4f5f-8a3c-edc02aff5325-serving-cert\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375740 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375763 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0774a9cb-d9cb-45a2-89e1-384902cf8f94-audit-policies\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375787 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04baa777-3a33-4b5a-9808-99a84b1b8005-config\") pod \"console-operator-58897d9998-slqc5\" (UID: \"04baa777-3a33-4b5a-9808-99a84b1b8005\") " pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375807 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/04baa777-3a33-4b5a-9808-99a84b1b8005-trusted-ca\") pod \"console-operator-58897d9998-slqc5\" (UID: \"04baa777-3a33-4b5a-9808-99a84b1b8005\") " pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375828 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-config\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375847 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-config\") pod \"route-controller-manager-6576b87f9c-5j7r4\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375868 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-config\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.375893 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04baa777-3a33-4b5a-9808-99a84b1b8005-serving-cert\") pod \"console-operator-58897d9998-slqc5\" (UID: \"04baa777-3a33-4b5a-9808-99a84b1b8005\") " pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.376176 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0774a9cb-d9cb-45a2-89e1-384902cf8f94-encryption-config\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.376235 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-serving-cert\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.376352 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.376832 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-audit\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.376884 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rpftn"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.377663 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.378592 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.378667 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4fbfc372-868f-41d5-81d0-efb8a3e88023-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4dn4m\" (UID: \"4fbfc372-868f-41d5-81d0-efb8a3e88023\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.378687 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.379001 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04baa777-3a33-4b5a-9808-99a84b1b8005-serving-cert\") pod \"console-operator-58897d9998-slqc5\" (UID: \"04baa777-3a33-4b5a-9808-99a84b1b8005\") " pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.379107 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-etcd-serving-ca\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.379222 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-config\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.379235 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0774a9cb-d9cb-45a2-89e1-384902cf8f94-audit-dir\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.379480 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-client-ca\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.379507 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6375157a-012b-441e-b97b-39d6830cef68-oauth-serving-cert\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.379669 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/744c22bb-2f35-494b-b352-ff71691bd0af-serving-cert\") pod \"openshift-config-operator-7777fb866f-pg4wt\" (UID: \"744c22bb-2f35-494b-b352-ff71691bd0af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.380281 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8-serviceca\") pod \"image-pruner-29563200-6l8ds\" (UID: \"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8\") " pod="openshift-image-registry/image-pruner-29563200-6l8ds" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.380390 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.380561 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fbfc372-868f-41d5-81d0-efb8a3e88023-config\") pod \"machine-api-operator-5694c8668f-4dn4m\" (UID: \"4fbfc372-868f-41d5-81d0-efb8a3e88023\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.380735 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.380838 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.380992 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.381098 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.381181 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6375157a-012b-441e-b97b-39d6830cef68-console-serving-cert\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.381601 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.382021 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-image-import-ca\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.382438 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79f09f2f-fe8b-4340-a1fc-c16f212b7e31-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2xj85\" (UID: \"79f09f2f-fe8b-4340-a1fc-c16f212b7e31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.382448 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0774a9cb-d9cb-45a2-89e1-384902cf8f94-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.382499 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/53bc5818-4983-4f5f-8a3c-edc02aff5325-node-pullsecrets\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.382755 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66e8121d-9422-4b23-93e9-9f4d350c149f-audit-dir\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.382860 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6375157a-012b-441e-b97b-39d6830cef68-service-ca\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.382911 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/53bc5818-4983-4f5f-8a3c-edc02aff5325-encryption-config\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.383484 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/651bef83-00db-40d1-b096-247b3c645859-config\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.383729 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/651bef83-00db-40d1-b096-247b3c645859-serving-cert\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.383766 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c447db7-c45e-4e01-925a-6ac56ec2aee2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-l46js\" (UID: \"5c447db7-c45e-4e01-925a-6ac56ec2aee2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.384387 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04baa777-3a33-4b5a-9808-99a84b1b8005-config\") pod \"console-operator-58897d9998-slqc5\" (UID: \"04baa777-3a33-4b5a-9808-99a84b1b8005\") " pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.384435 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/651bef83-00db-40d1-b096-247b3c645859-etcd-service-ca\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.384500 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.384598 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0774a9cb-d9cb-45a2-89e1-384902cf8f94-audit-policies\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.384717 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.384776 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/04baa777-3a33-4b5a-9808-99a84b1b8005-trusted-ca\") pod \"console-operator-58897d9998-slqc5\" (UID: \"04baa777-3a33-4b5a-9808-99a84b1b8005\") " pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.385317 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6375157a-012b-441e-b97b-39d6830cef68-trusted-ca-bundle\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.385377 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.385395 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.385375 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53bc5818-4983-4f5f-8a3c-edc02aff5325-config\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.385758 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-config\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.385837 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-config\") pod \"route-controller-manager-6576b87f9c-5j7r4\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.386173 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c447db7-c45e-4e01-925a-6ac56ec2aee2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-l46js\" (UID: \"5c447db7-c45e-4e01-925a-6ac56ec2aee2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.386236 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b2013fc-d64f-4f21-be62-e17f3855f6c1-serving-cert\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.386335 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6375157a-012b-441e-b97b-39d6830cef68-console-oauth-config\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.386748 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6e69d708-f2df-436e-a5ec-41848aecbace-metrics-tls\") pod \"dns-operator-744455d44c-w4s69\" (UID: \"6e69d708-f2df-436e-a5ec-41848aecbace\") " pod="openshift-dns-operator/dns-operator-744455d44c-w4s69" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.386851 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-serving-cert\") pod \"route-controller-manager-6576b87f9c-5j7r4\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.386957 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-audit-policies\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.387053 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.387265 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.387601 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.387840 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79f09f2f-fe8b-4340-a1fc-c16f212b7e31-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2xj85\" (UID: \"79f09f2f-fe8b-4340-a1fc-c16f212b7e31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.388032 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.388242 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0774a9cb-d9cb-45a2-89e1-384902cf8f94-etcd-client\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.388319 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53bc5818-4983-4f5f-8a3c-edc02aff5325-serving-cert\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.389002 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.389356 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.389406 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.389708 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.390383 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0774a9cb-d9cb-45a2-89e1-384902cf8f94-serving-cert\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.390589 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/651bef83-00db-40d1-b096-247b3c645859-etcd-client\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.390710 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/812c885e-98f8-47cb-afd6-5fb537bea3e3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-w8wsj\" (UID: \"812c885e-98f8-47cb-afd6-5fb537bea3e3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.391683 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563210-lptd5"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.392675 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/53bc5818-4983-4f5f-8a3c-edc02aff5325-etcd-client\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.392994 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.394166 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.395293 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-f8svz"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.396427 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.397505 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b8gtz"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.398693 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gx8mh"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.400197 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qkkt2"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.401946 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-85cfx"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.407403 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-k424z"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.408055 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-kzks7"] Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.418004 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.428762 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.449430 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.469120 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.476497 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8qn5\" (UniqueName: \"kubernetes.io/projected/b876a42a-856b-4530-a601-3ca35db1078a-kube-api-access-v8qn5\") pod \"migrator-59844c95c7-wxstd\" (UID: \"b876a42a-856b-4530-a601-3ca35db1078a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wxstd" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.489978 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.509809 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.529996 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.550196 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.569514 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.602340 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.609264 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.630234 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.649642 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.670753 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.678376 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.689579 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.710613 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.729702 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.749446 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.769583 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.789799 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.809912 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.830875 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.858907 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.870116 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.891441 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.910331 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.930395 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.949914 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.969786 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 18 00:11:43 crc kubenswrapper[4983]: I0318 00:11:43.989745 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.009970 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.030891 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.050818 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.071049 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.089756 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.131043 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.149804 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.170811 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.189291 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.210211 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.229880 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.250818 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.270749 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.289862 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.310084 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.328506 4983 request.go:700] Waited for 1.013906927s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/secrets?fieldSelector=metadata.name%3Dservice-ca-operator-dockercfg-rg9jl&limit=500&resourceVersion=0 Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.329866 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.350318 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.371002 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.389976 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.410640 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.431617 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.451230 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.470968 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.490880 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.510181 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.530183 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.549438 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.570192 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.589619 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.609602 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.634874 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.650301 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.670226 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.690826 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.731207 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.749895 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.770397 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.789920 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.809471 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.830082 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.851177 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.870498 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.890478 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.910603 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.930389 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.950523 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.970594 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 18 00:11:44 crc kubenswrapper[4983]: I0318 00:11:44.990656 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.010178 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.031058 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.050586 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.070965 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.089992 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.111045 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.130425 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.150552 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.170117 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.190486 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.210193 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.230648 4983 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.249606 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.297345 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq7j2\" (UniqueName: \"kubernetes.io/projected/0b2013fc-d64f-4f21-be62-e17f3855f6c1-kube-api-access-lq7j2\") pod \"controller-manager-879f6c89f-zlqph\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.317830 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqcc9\" (UniqueName: \"kubernetes.io/projected/6e69d708-f2df-436e-a5ec-41848aecbace-kube-api-access-zqcc9\") pod \"dns-operator-744455d44c-w4s69\" (UID: \"6e69d708-f2df-436e-a5ec-41848aecbace\") " pod="openshift-dns-operator/dns-operator-744455d44c-w4s69" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.328626 4983 request.go:700] Waited for 1.953098715s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/serviceaccounts/route-controller-manager-sa/token Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.339780 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2ms6\" (UniqueName: \"kubernetes.io/projected/c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8-kube-api-access-t2ms6\") pod \"image-pruner-29563200-6l8ds\" (UID: \"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8\") " pod="openshift-image-registry/image-pruner-29563200-6l8ds" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.359100 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf67f\" (UniqueName: \"kubernetes.io/projected/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-kube-api-access-nf67f\") pod \"route-controller-manager-6576b87f9c-5j7r4\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.367092 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29563200-6l8ds" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.380841 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqwc6\" (UniqueName: \"kubernetes.io/projected/e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8-kube-api-access-vqwc6\") pod \"authentication-operator-69f744f599-z7h4m\" (UID: \"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.399675 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-w4s69" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.403321 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kcsp\" (UniqueName: \"kubernetes.io/projected/6375157a-012b-441e-b97b-39d6830cef68-kube-api-access-5kcsp\") pod \"console-f9d7485db-65l25\" (UID: \"6375157a-012b-441e-b97b-39d6830cef68\") " pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.421108 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pkkq\" (UniqueName: \"kubernetes.io/projected/744c22bb-2f35-494b-b352-ff71691bd0af-kube-api-access-6pkkq\") pod \"openshift-config-operator-7777fb866f-pg4wt\" (UID: \"744c22bb-2f35-494b-b352-ff71691bd0af\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.429889 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx7hh\" (UniqueName: \"kubernetes.io/projected/79f09f2f-fe8b-4340-a1fc-c16f212b7e31-kube-api-access-gx7hh\") pod \"openshift-controller-manager-operator-756b6f6bc6-2xj85\" (UID: \"79f09f2f-fe8b-4340-a1fc-c16f212b7e31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.433727 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.441967 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.458635 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tgpb\" (UniqueName: \"kubernetes.io/projected/812c885e-98f8-47cb-afd6-5fb537bea3e3-kube-api-access-8tgpb\") pod \"cluster-samples-operator-665b6dd947-w8wsj\" (UID: \"812c885e-98f8-47cb-afd6-5fb537bea3e3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.471026 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.477449 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2p82\" (UniqueName: \"kubernetes.io/projected/66e8121d-9422-4b23-93e9-9f4d350c149f-kube-api-access-m2p82\") pod \"oauth-openshift-558db77b4-6t6lv\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.477697 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.496754 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxm5m\" (UniqueName: \"kubernetes.io/projected/4fbfc372-868f-41d5-81d0-efb8a3e88023-kube-api-access-dxm5m\") pod \"machine-api-operator-5694c8668f-4dn4m\" (UID: \"4fbfc372-868f-41d5-81d0-efb8a3e88023\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.525782 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqntz\" (UniqueName: \"kubernetes.io/projected/651bef83-00db-40d1-b096-247b3c645859-kube-api-access-wqntz\") pod \"etcd-operator-b45778765-gvd97\" (UID: \"651bef83-00db-40d1-b096-247b3c645859\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.528596 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-246g7\" (UniqueName: \"kubernetes.io/projected/53bc5818-4983-4f5f-8a3c-edc02aff5325-kube-api-access-246g7\") pod \"apiserver-76f77b778f-h74lp\" (UID: \"53bc5818-4983-4f5f-8a3c-edc02aff5325\") " pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.547315 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvvzr\" (UniqueName: \"kubernetes.io/projected/04baa777-3a33-4b5a-9808-99a84b1b8005-kube-api-access-lvvzr\") pod \"console-operator-58897d9998-slqc5\" (UID: \"04baa777-3a33-4b5a-9808-99a84b1b8005\") " pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.565342 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4887x\" (UniqueName: \"kubernetes.io/projected/5c447db7-c45e-4e01-925a-6ac56ec2aee2-kube-api-access-4887x\") pod \"openshift-apiserver-operator-796bbdcf4f-l46js\" (UID: \"5c447db7-c45e-4e01-925a-6ac56ec2aee2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.571915 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.584381 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.597202 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29563200-6l8ds"] Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.602105 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.608504 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tpgg\" (UniqueName: \"kubernetes.io/projected/0774a9cb-d9cb-45a2-89e1-384902cf8f94-kube-api-access-4tpgg\") pod \"apiserver-7bbb656c7d-g7fq4\" (UID: \"0774a9cb-d9cb-45a2-89e1-384902cf8f94\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.613333 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8qn5\" (UniqueName: \"kubernetes.io/projected/b876a42a-856b-4530-a601-3ca35db1078a-kube-api-access-v8qn5\") pod \"migrator-59844c95c7-wxstd\" (UID: \"b876a42a-856b-4530-a601-3ca35db1078a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wxstd" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.618512 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.618608 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.668751 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.708886 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a57b48db-a186-4a73-b139-a9249a19a3ed-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zsklf\" (UID: \"a57b48db-a186-4a73-b139-a9249a19a3ed\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.708977 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/297c1426-c499-449c-ab44-e1f898262839-registry-certificates\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709024 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/49880d65-bff4-45d1-8fb7-907e524568fc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-99dq7\" (UID: \"49880d65-bff4-45d1-8fb7-907e524568fc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709047 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvpqz\" (UniqueName: \"kubernetes.io/projected/eec36046-1daf-4f76-9a9e-3d9001830bb2-kube-api-access-hvpqz\") pod \"cluster-image-registry-operator-dc59b4c8b-hmcxg\" (UID: \"eec36046-1daf-4f76-9a9e-3d9001830bb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709102 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eec36046-1daf-4f76-9a9e-3d9001830bb2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hmcxg\" (UID: \"eec36046-1daf-4f76-9a9e-3d9001830bb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709119 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71e981d5-d657-46ef-9ec0-4ddad58be6ee-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rpftn\" (UID: \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\") " pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709155 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/297c1426-c499-449c-ab44-e1f898262839-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709189 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr5cg\" (UniqueName: \"kubernetes.io/projected/beb8a379-4733-45b5-b31b-0e02eb9d5649-kube-api-access-sr5cg\") pod \"machine-approver-56656f9798-44428\" (UID: \"beb8a379-4733-45b5-b31b-0e02eb9d5649\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709223 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6d53acd4-182c-4dd4-93e1-bccac2a0a522-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-85cfx\" (UID: \"6d53acd4-182c-4dd4-93e1-bccac2a0a522\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-85cfx" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709256 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7drlm\" (UniqueName: \"kubernetes.io/projected/dafd0e48-880c-44fc-b8c5-8056a37dac27-kube-api-access-7drlm\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709325 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dafd0e48-880c-44fc-b8c5-8056a37dac27-metrics-certs\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709347 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/dafd0e48-880c-44fc-b8c5-8056a37dac27-stats-auth\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709399 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709420 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rrqj\" (UniqueName: \"kubernetes.io/projected/a57b48db-a186-4a73-b139-a9249a19a3ed-kube-api-access-7rrqj\") pod \"ingress-operator-5b745b69d9-zsklf\" (UID: \"a57b48db-a186-4a73-b139-a9249a19a3ed\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709497 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nclc8\" (UniqueName: \"kubernetes.io/projected/71e981d5-d657-46ef-9ec0-4ddad58be6ee-kube-api-access-nclc8\") pod \"marketplace-operator-79b997595-rpftn\" (UID: \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\") " pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709528 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq65z\" (UniqueName: \"kubernetes.io/projected/6d53acd4-182c-4dd4-93e1-bccac2a0a522-kube-api-access-hq65z\") pod \"multus-admission-controller-857f4d67dd-85cfx\" (UID: \"6d53acd4-182c-4dd4-93e1-bccac2a0a522\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-85cfx" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709573 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-registry-tls\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709604 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/297c1426-c499-449c-ab44-e1f898262839-trusted-ca\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709622 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/eec36046-1daf-4f76-9a9e-3d9001830bb2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hmcxg\" (UID: \"eec36046-1daf-4f76-9a9e-3d9001830bb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709640 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/eaebd031-fb05-489b-8be1-e4e115bdcb7d-images\") pod \"machine-config-operator-74547568cd-9h7fc\" (UID: \"eaebd031-fb05-489b-8be1-e4e115bdcb7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709706 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a57b48db-a186-4a73-b139-a9249a19a3ed-metrics-tls\") pod \"ingress-operator-5b745b69d9-zsklf\" (UID: \"a57b48db-a186-4a73-b139-a9249a19a3ed\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709798 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eaebd031-fb05-489b-8be1-e4e115bdcb7d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9h7fc\" (UID: \"eaebd031-fb05-489b-8be1-e4e115bdcb7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709828 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/49880d65-bff4-45d1-8fb7-907e524568fc-proxy-tls\") pod \"machine-config-controller-84d6567774-99dq7\" (UID: \"49880d65-bff4-45d1-8fb7-907e524568fc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709850 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eec36046-1daf-4f76-9a9e-3d9001830bb2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hmcxg\" (UID: \"eec36046-1daf-4f76-9a9e-3d9001830bb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709887 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/297c1426-c499-449c-ab44-e1f898262839-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709948 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/beb8a379-4733-45b5-b31b-0e02eb9d5649-auth-proxy-config\") pod \"machine-approver-56656f9798-44428\" (UID: \"beb8a379-4733-45b5-b31b-0e02eb9d5649\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.709987 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-bound-sa-token\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.715045 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqrff\" (UniqueName: \"kubernetes.io/projected/a4e42b8f-7120-4b47-b8ed-725640893464-kube-api-access-lqrff\") pod \"control-plane-machine-set-operator-78cbb6b69f-q72ss\" (UID: \"a4e42b8f-7120-4b47-b8ed-725640893464\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss" Mar 18 00:11:45 crc kubenswrapper[4983]: E0318 00:11:45.715341 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:46.21507797 +0000 UTC m=+224.112452640 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.715669 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.716586 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a4e42b8f-7120-4b47-b8ed-725640893464-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-q72ss\" (UID: \"a4e42b8f-7120-4b47-b8ed-725640893464\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.716795 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a57b48db-a186-4a73-b139-a9249a19a3ed-trusted-ca\") pod \"ingress-operator-5b745b69d9-zsklf\" (UID: \"a57b48db-a186-4a73-b139-a9249a19a3ed\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.716876 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dafd0e48-880c-44fc-b8c5-8056a37dac27-service-ca-bundle\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.717299 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/eaebd031-fb05-489b-8be1-e4e115bdcb7d-proxy-tls\") pod \"machine-config-operator-74547568cd-9h7fc\" (UID: \"eaebd031-fb05-489b-8be1-e4e115bdcb7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.717319 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68k9x\" (UniqueName: \"kubernetes.io/projected/eb9a12c3-1119-4b1e-8d32-0a9696a00392-kube-api-access-68k9x\") pod \"downloads-7954f5f757-87bh5\" (UID: \"eb9a12c3-1119-4b1e-8d32-0a9696a00392\") " pod="openshift-console/downloads-7954f5f757-87bh5" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.717717 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6grh\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-kube-api-access-l6grh\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.718850 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/71e981d5-d657-46ef-9ec0-4ddad58be6ee-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rpftn\" (UID: \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\") " pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.719169 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/dafd0e48-880c-44fc-b8c5-8056a37dac27-default-certificate\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.719337 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln6dp\" (UniqueName: \"kubernetes.io/projected/49880d65-bff4-45d1-8fb7-907e524568fc-kube-api-access-ln6dp\") pod \"machine-config-controller-84d6567774-99dq7\" (UID: \"49880d65-bff4-45d1-8fb7-907e524568fc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.719527 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/beb8a379-4733-45b5-b31b-0e02eb9d5649-machine-approver-tls\") pod \"machine-approver-56656f9798-44428\" (UID: \"beb8a379-4733-45b5-b31b-0e02eb9d5649\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.719550 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beb8a379-4733-45b5-b31b-0e02eb9d5649-config\") pod \"machine-approver-56656f9798-44428\" (UID: \"beb8a379-4733-45b5-b31b-0e02eb9d5649\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.719593 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c77mr\" (UniqueName: \"kubernetes.io/projected/eaebd031-fb05-489b-8be1-e4e115bdcb7d-kube-api-access-c77mr\") pod \"machine-config-operator-74547568cd-9h7fc\" (UID: \"eaebd031-fb05-489b-8be1-e4e115bdcb7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.723065 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.761615 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.823069 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:45 crc kubenswrapper[4983]: E0318 00:11:45.823409 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:46.323376392 +0000 UTC m=+224.220751002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826252 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/297c1426-c499-449c-ab44-e1f898262839-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826283 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eec36046-1daf-4f76-9a9e-3d9001830bb2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hmcxg\" (UID: \"eec36046-1daf-4f76-9a9e-3d9001830bb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826306 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71e981d5-d657-46ef-9ec0-4ddad58be6ee-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rpftn\" (UID: \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\") " pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826325 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr5cg\" (UniqueName: \"kubernetes.io/projected/beb8a379-4733-45b5-b31b-0e02eb9d5649-kube-api-access-sr5cg\") pod \"machine-approver-56656f9798-44428\" (UID: \"beb8a379-4733-45b5-b31b-0e02eb9d5649\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826346 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a-config\") pod \"service-ca-operator-777779d784-f8svz\" (UID: \"f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826362 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/602912b1-d12a-406a-ba6e-e0d94dc62038-config-volume\") pod \"collect-profiles-29563200-xtxcb\" (UID: \"602912b1-d12a-406a-ba6e-e0d94dc62038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826385 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6d53acd4-182c-4dd4-93e1-bccac2a0a522-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-85cfx\" (UID: \"6d53acd4-182c-4dd4-93e1-bccac2a0a522\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-85cfx" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826415 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7drlm\" (UniqueName: \"kubernetes.io/projected/dafd0e48-880c-44fc-b8c5-8056a37dac27-kube-api-access-7drlm\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826440 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd950f1d-4d97-4ed4-8d9c-04245019bbdb-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kf9gt\" (UID: \"bd950f1d-4d97-4ed4-8d9c-04245019bbdb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826468 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dafd0e48-880c-44fc-b8c5-8056a37dac27-metrics-certs\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826485 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jslb\" (UID: \"79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826499 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-csi-data-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826516 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/dafd0e48-880c-44fc-b8c5-8056a37dac27-stats-auth\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826536 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29fc0965-6c9b-455f-b83e-f2fd1e04c6dc-config-volume\") pod \"dns-default-k424z\" (UID: \"29fc0965-6c9b-455f-b83e-f2fd1e04c6dc\") " pod="openshift-dns/dns-default-k424z" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826562 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826578 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rrqj\" (UniqueName: \"kubernetes.io/projected/a57b48db-a186-4a73-b139-a9249a19a3ed-kube-api-access-7rrqj\") pod \"ingress-operator-5b745b69d9-zsklf\" (UID: \"a57b48db-a186-4a73-b139-a9249a19a3ed\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826593 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-plugins-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826611 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9x64\" (UniqueName: \"kubernetes.io/projected/799072d7-3ce6-4742-be4c-065ecea7e24c-kube-api-access-h9x64\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826637 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nclc8\" (UniqueName: \"kubernetes.io/projected/71e981d5-d657-46ef-9ec0-4ddad58be6ee-kube-api-access-nclc8\") pod \"marketplace-operator-79b997595-rpftn\" (UID: \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\") " pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826655 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-registry-tls\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826671 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq65z\" (UniqueName: \"kubernetes.io/projected/6d53acd4-182c-4dd4-93e1-bccac2a0a522-kube-api-access-hq65z\") pod \"multus-admission-controller-857f4d67dd-85cfx\" (UID: \"6d53acd4-182c-4dd4-93e1-bccac2a0a522\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-85cfx" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826689 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/297c1426-c499-449c-ab44-e1f898262839-trusted-ca\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826705 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84bde919-6dad-4080-977f-81cda3119a89-config\") pod \"kube-controller-manager-operator-78b949d7b-w4hpj\" (UID: \"84bde919-6dad-4080-977f-81cda3119a89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826722 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-socket-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826738 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7ddf406d-94c4-4ce9-b593-aebe0fca2b51-srv-cert\") pod \"olm-operator-6b444d44fb-8qf4k\" (UID: \"7ddf406d-94c4-4ce9-b593-aebe0fca2b51\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826752 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8fvq\" (UniqueName: \"kubernetes.io/projected/29fc0965-6c9b-455f-b83e-f2fd1e04c6dc-kube-api-access-v8fvq\") pod \"dns-default-k424z\" (UID: \"29fc0965-6c9b-455f-b83e-f2fd1e04c6dc\") " pod="openshift-dns/dns-default-k424z" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826779 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/eec36046-1daf-4f76-9a9e-3d9001830bb2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hmcxg\" (UID: \"eec36046-1daf-4f76-9a9e-3d9001830bb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826796 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/eaebd031-fb05-489b-8be1-e4e115bdcb7d-images\") pod \"machine-config-operator-74547568cd-9h7fc\" (UID: \"eaebd031-fb05-489b-8be1-e4e115bdcb7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826813 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a57b48db-a186-4a73-b139-a9249a19a3ed-metrics-tls\") pod \"ingress-operator-5b745b69d9-zsklf\" (UID: \"a57b48db-a186-4a73-b139-a9249a19a3ed\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826831 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjvwm\" (UniqueName: \"kubernetes.io/projected/d3f900da-08ff-4983-89b7-4dea2613f0fd-kube-api-access-wjvwm\") pod \"service-ca-9c57cc56f-gx8mh\" (UID: \"d3f900da-08ff-4983-89b7-4dea2613f0fd\") " pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826847 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8-tmpfs\") pod \"packageserver-d55dfcdfc-9xkhq\" (UID: \"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826864 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jslb\" (UID: \"79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826880 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ecc96f36-7ea7-4beb-977b-ae803b665b16-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kqxb6\" (UID: \"ecc96f36-7ea7-4beb-977b-ae803b665b16\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826901 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eaebd031-fb05-489b-8be1-e4e115bdcb7d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9h7fc\" (UID: \"eaebd031-fb05-489b-8be1-e4e115bdcb7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826920 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a-serving-cert\") pod \"service-ca-operator-777779d784-f8svz\" (UID: \"f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826961 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8fjh\" (UniqueName: \"kubernetes.io/projected/bd950f1d-4d97-4ed4-8d9c-04245019bbdb-kube-api-access-t8fjh\") pod \"kube-storage-version-migrator-operator-b67b599dd-kf9gt\" (UID: \"bd950f1d-4d97-4ed4-8d9c-04245019bbdb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826983 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecc96f36-7ea7-4beb-977b-ae803b665b16-config\") pod \"kube-apiserver-operator-766d6c64bb-kqxb6\" (UID: \"ecc96f36-7ea7-4beb-977b-ae803b665b16\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827011 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/49880d65-bff4-45d1-8fb7-907e524568fc-proxy-tls\") pod \"machine-config-controller-84d6567774-99dq7\" (UID: \"49880d65-bff4-45d1-8fb7-907e524568fc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827034 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eec36046-1daf-4f76-9a9e-3d9001830bb2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hmcxg\" (UID: \"eec36046-1daf-4f76-9a9e-3d9001830bb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827058 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f9476891-3ac9-4914-9261-7d82a3367a00-certs\") pod \"machine-config-server-pch42\" (UID: \"f9476891-3ac9-4914-9261-7d82a3367a00\") " pod="openshift-machine-config-operator/machine-config-server-pch42" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827080 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84bde919-6dad-4080-977f-81cda3119a89-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-w4hpj\" (UID: \"84bde919-6dad-4080-977f-81cda3119a89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827109 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/297c1426-c499-449c-ab44-e1f898262839-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827132 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8-apiservice-cert\") pod \"packageserver-d55dfcdfc-9xkhq\" (UID: \"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827161 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1321ddb2-5769-482f-b4dc-1971dad30fdf-profile-collector-cert\") pod \"catalog-operator-68c6474976-wsk9g\" (UID: \"1321ddb2-5769-482f-b4dc-1971dad30fdf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827189 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/beb8a379-4733-45b5-b31b-0e02eb9d5649-auth-proxy-config\") pod \"machine-approver-56656f9798-44428\" (UID: \"beb8a379-4733-45b5-b31b-0e02eb9d5649\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827213 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd950f1d-4d97-4ed4-8d9c-04245019bbdb-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kf9gt\" (UID: \"bd950f1d-4d97-4ed4-8d9c-04245019bbdb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827243 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-bound-sa-token\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827267 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqrff\" (UniqueName: \"kubernetes.io/projected/a4e42b8f-7120-4b47-b8ed-725640893464-kube-api-access-lqrff\") pod \"control-plane-machine-set-operator-78cbb6b69f-q72ss\" (UID: \"a4e42b8f-7120-4b47-b8ed-725640893464\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827298 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5qrv\" (UniqueName: \"kubernetes.io/projected/63bc8988-a69a-4385-8ff3-45edafde495a-kube-api-access-p5qrv\") pod \"ingress-canary-b8gtz\" (UID: \"63bc8988-a69a-4385-8ff3-45edafde495a\") " pod="openshift-ingress-canary/ingress-canary-b8gtz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827318 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlv5c\" (UniqueName: \"kubernetes.io/projected/da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8-kube-api-access-nlv5c\") pod \"packageserver-d55dfcdfc-9xkhq\" (UID: \"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827340 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a57b48db-a186-4a73-b139-a9249a19a3ed-trusted-ca\") pod \"ingress-operator-5b745b69d9-zsklf\" (UID: \"a57b48db-a186-4a73-b139-a9249a19a3ed\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827368 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a4e42b8f-7120-4b47-b8ed-725640893464-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-q72ss\" (UID: \"a4e42b8f-7120-4b47-b8ed-725640893464\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827394 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-mountpoint-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827417 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dafd0e48-880c-44fc-b8c5-8056a37dac27-service-ca-bundle\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827434 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jslb\" (UID: \"79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827449 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7237321-db39-4ed6-a22e-b3b57c5b6be6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-96hcz\" (UID: \"d7237321-db39-4ed6-a22e-b3b57c5b6be6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827467 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r824j\" (UniqueName: \"kubernetes.io/projected/602912b1-d12a-406a-ba6e-e0d94dc62038-kube-api-access-r824j\") pod \"collect-profiles-29563200-xtxcb\" (UID: \"602912b1-d12a-406a-ba6e-e0d94dc62038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827490 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7ddf406d-94c4-4ce9-b593-aebe0fca2b51-profile-collector-cert\") pod \"olm-operator-6b444d44fb-8qf4k\" (UID: \"7ddf406d-94c4-4ce9-b593-aebe0fca2b51\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827510 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/eaebd031-fb05-489b-8be1-e4e115bdcb7d-proxy-tls\") pod \"machine-config-operator-74547568cd-9h7fc\" (UID: \"eaebd031-fb05-489b-8be1-e4e115bdcb7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827528 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68k9x\" (UniqueName: \"kubernetes.io/projected/eb9a12c3-1119-4b1e-8d32-0a9696a00392-kube-api-access-68k9x\") pod \"downloads-7954f5f757-87bh5\" (UID: \"eb9a12c3-1119-4b1e-8d32-0a9696a00392\") " pod="openshift-console/downloads-7954f5f757-87bh5" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827547 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6grh\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-kube-api-access-l6grh\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827564 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/602912b1-d12a-406a-ba6e-e0d94dc62038-secret-volume\") pod \"collect-profiles-29563200-xtxcb\" (UID: \"602912b1-d12a-406a-ba6e-e0d94dc62038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827592 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rf9n\" (UniqueName: \"kubernetes.io/projected/f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a-kube-api-access-6rf9n\") pod \"service-ca-operator-777779d784-f8svz\" (UID: \"f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827608 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84bde919-6dad-4080-977f-81cda3119a89-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-w4hpj\" (UID: \"84bde919-6dad-4080-977f-81cda3119a89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827647 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/71e981d5-d657-46ef-9ec0-4ddad58be6ee-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rpftn\" (UID: \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\") " pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827667 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/dafd0e48-880c-44fc-b8c5-8056a37dac27-default-certificate\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827683 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/63bc8988-a69a-4385-8ff3-45edafde495a-cert\") pod \"ingress-canary-b8gtz\" (UID: \"63bc8988-a69a-4385-8ff3-45edafde495a\") " pod="openshift-ingress-canary/ingress-canary-b8gtz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827700 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kplv\" (UniqueName: \"kubernetes.io/projected/efeae5ef-dba6-4031-83be-8f493fe853b9-kube-api-access-5kplv\") pod \"auto-csr-approver-29563210-lptd5\" (UID: \"efeae5ef-dba6-4031-83be-8f493fe853b9\") " pod="openshift-infra/auto-csr-approver-29563210-lptd5" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827721 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln6dp\" (UniqueName: \"kubernetes.io/projected/49880d65-bff4-45d1-8fb7-907e524568fc-kube-api-access-ln6dp\") pod \"machine-config-controller-84d6567774-99dq7\" (UID: \"49880d65-bff4-45d1-8fb7-907e524568fc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827743 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/beb8a379-4733-45b5-b31b-0e02eb9d5649-machine-approver-tls\") pod \"machine-approver-56656f9798-44428\" (UID: \"beb8a379-4733-45b5-b31b-0e02eb9d5649\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827759 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beb8a379-4733-45b5-b31b-0e02eb9d5649-config\") pod \"machine-approver-56656f9798-44428\" (UID: \"beb8a379-4733-45b5-b31b-0e02eb9d5649\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827775 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-registration-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827792 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c77mr\" (UniqueName: \"kubernetes.io/projected/eaebd031-fb05-489b-8be1-e4e115bdcb7d-kube-api-access-c77mr\") pod \"machine-config-operator-74547568cd-9h7fc\" (UID: \"eaebd031-fb05-489b-8be1-e4e115bdcb7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827810 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtxh2\" (UniqueName: \"kubernetes.io/projected/f9476891-3ac9-4914-9261-7d82a3367a00-kube-api-access-mtxh2\") pod \"machine-config-server-pch42\" (UID: \"f9476891-3ac9-4914-9261-7d82a3367a00\") " pod="openshift-machine-config-operator/machine-config-server-pch42" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827826 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj9pm\" (UniqueName: \"kubernetes.io/projected/7ddf406d-94c4-4ce9-b593-aebe0fca2b51-kube-api-access-cj9pm\") pod \"olm-operator-6b444d44fb-8qf4k\" (UID: \"7ddf406d-94c4-4ce9-b593-aebe0fca2b51\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827844 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8-webhook-cert\") pod \"packageserver-d55dfcdfc-9xkhq\" (UID: \"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827858 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1321ddb2-5769-482f-b4dc-1971dad30fdf-srv-cert\") pod \"catalog-operator-68c6474976-wsk9g\" (UID: \"1321ddb2-5769-482f-b4dc-1971dad30fdf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827873 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cs7v\" (UniqueName: \"kubernetes.io/projected/1321ddb2-5769-482f-b4dc-1971dad30fdf-kube-api-access-9cs7v\") pod \"catalog-operator-68c6474976-wsk9g\" (UID: \"1321ddb2-5769-482f-b4dc-1971dad30fdf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827889 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d3f900da-08ff-4983-89b7-4dea2613f0fd-signing-cabundle\") pod \"service-ca-9c57cc56f-gx8mh\" (UID: \"d3f900da-08ff-4983-89b7-4dea2613f0fd\") " pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827904 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f9476891-3ac9-4914-9261-7d82a3367a00-node-bootstrap-token\") pod \"machine-config-server-pch42\" (UID: \"f9476891-3ac9-4914-9261-7d82a3367a00\") " pod="openshift-machine-config-operator/machine-config-server-pch42" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.827919 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq2gd\" (UniqueName: \"kubernetes.io/projected/d7237321-db39-4ed6-a22e-b3b57c5b6be6-kube-api-access-xq2gd\") pod \"package-server-manager-789f6589d5-96hcz\" (UID: \"d7237321-db39-4ed6-a22e-b3b57c5b6be6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.828061 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d3f900da-08ff-4983-89b7-4dea2613f0fd-signing-key\") pod \"service-ca-9c57cc56f-gx8mh\" (UID: \"d3f900da-08ff-4983-89b7-4dea2613f0fd\") " pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.828121 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a57b48db-a186-4a73-b139-a9249a19a3ed-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zsklf\" (UID: \"a57b48db-a186-4a73-b139-a9249a19a3ed\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.828148 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/297c1426-c499-449c-ab44-e1f898262839-registry-certificates\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.828178 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ecc96f36-7ea7-4beb-977b-ae803b665b16-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kqxb6\" (UID: \"ecc96f36-7ea7-4beb-977b-ae803b665b16\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.828212 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/49880d65-bff4-45d1-8fb7-907e524568fc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-99dq7\" (UID: \"49880d65-bff4-45d1-8fb7-907e524568fc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.828238 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvpqz\" (UniqueName: \"kubernetes.io/projected/eec36046-1daf-4f76-9a9e-3d9001830bb2-kube-api-access-hvpqz\") pod \"cluster-image-registry-operator-dc59b4c8b-hmcxg\" (UID: \"eec36046-1daf-4f76-9a9e-3d9001830bb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.828266 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/29fc0965-6c9b-455f-b83e-f2fd1e04c6dc-metrics-tls\") pod \"dns-default-k424z\" (UID: \"29fc0965-6c9b-455f-b83e-f2fd1e04c6dc\") " pod="openshift-dns/dns-default-k424z" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.829361 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dafd0e48-880c-44fc-b8c5-8056a37dac27-service-ca-bundle\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.830631 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71e981d5-d657-46ef-9ec0-4ddad58be6ee-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rpftn\" (UID: \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\") " pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.826790 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/297c1426-c499-449c-ab44-e1f898262839-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.831998 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/beb8a379-4733-45b5-b31b-0e02eb9d5649-auth-proxy-config\") pod \"machine-approver-56656f9798-44428\" (UID: \"beb8a379-4733-45b5-b31b-0e02eb9d5649\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.832857 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/297c1426-c499-449c-ab44-e1f898262839-trusted-ca\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: E0318 00:11:45.833331 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:46.333309808 +0000 UTC m=+224.230684418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.833896 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-registry-tls\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.834013 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beb8a379-4733-45b5-b31b-0e02eb9d5649-config\") pod \"machine-approver-56656f9798-44428\" (UID: \"beb8a379-4733-45b5-b31b-0e02eb9d5649\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.836438 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/297c1426-c499-449c-ab44-e1f898262839-registry-certificates\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.837305 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6d53acd4-182c-4dd4-93e1-bccac2a0a522-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-85cfx\" (UID: \"6d53acd4-182c-4dd4-93e1-bccac2a0a522\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-85cfx" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.837388 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/beb8a379-4733-45b5-b31b-0e02eb9d5649-machine-approver-tls\") pod \"machine-approver-56656f9798-44428\" (UID: \"beb8a379-4733-45b5-b31b-0e02eb9d5649\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.838113 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eaebd031-fb05-489b-8be1-e4e115bdcb7d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9h7fc\" (UID: \"eaebd031-fb05-489b-8be1-e4e115bdcb7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.838223 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dafd0e48-880c-44fc-b8c5-8056a37dac27-metrics-certs\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.838520 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/eec36046-1daf-4f76-9a9e-3d9001830bb2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hmcxg\" (UID: \"eec36046-1daf-4f76-9a9e-3d9001830bb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.838723 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eec36046-1daf-4f76-9a9e-3d9001830bb2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hmcxg\" (UID: \"eec36046-1daf-4f76-9a9e-3d9001830bb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.838752 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/dafd0e48-880c-44fc-b8c5-8056a37dac27-default-certificate\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.838813 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a57b48db-a186-4a73-b139-a9249a19a3ed-trusted-ca\") pod \"ingress-operator-5b745b69d9-zsklf\" (UID: \"a57b48db-a186-4a73-b139-a9249a19a3ed\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.839801 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/49880d65-bff4-45d1-8fb7-907e524568fc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-99dq7\" (UID: \"49880d65-bff4-45d1-8fb7-907e524568fc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.840804 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a4e42b8f-7120-4b47-b8ed-725640893464-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-q72ss\" (UID: \"a4e42b8f-7120-4b47-b8ed-725640893464\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.840849 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/eaebd031-fb05-489b-8be1-e4e115bdcb7d-images\") pod \"machine-config-operator-74547568cd-9h7fc\" (UID: \"eaebd031-fb05-489b-8be1-e4e115bdcb7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.841235 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.841241 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/dafd0e48-880c-44fc-b8c5-8056a37dac27-stats-auth\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.841547 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/297c1426-c499-449c-ab44-e1f898262839-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.843244 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a57b48db-a186-4a73-b139-a9249a19a3ed-metrics-tls\") pod \"ingress-operator-5b745b69d9-zsklf\" (UID: \"a57b48db-a186-4a73-b139-a9249a19a3ed\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.843608 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/71e981d5-d657-46ef-9ec0-4ddad58be6ee-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rpftn\" (UID: \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\") " pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.843636 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/eaebd031-fb05-489b-8be1-e4e115bdcb7d-proxy-tls\") pod \"machine-config-operator-74547568cd-9h7fc\" (UID: \"eaebd031-fb05-489b-8be1-e4e115bdcb7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.847299 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wxstd" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.848124 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/49880d65-bff4-45d1-8fb7-907e524568fc-proxy-tls\") pod \"machine-config-controller-84d6567774-99dq7\" (UID: \"49880d65-bff4-45d1-8fb7-907e524568fc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.867101 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rrqj\" (UniqueName: \"kubernetes.io/projected/a57b48db-a186-4a73-b139-a9249a19a3ed-kube-api-access-7rrqj\") pod \"ingress-operator-5b745b69d9-zsklf\" (UID: \"a57b48db-a186-4a73-b139-a9249a19a3ed\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.907591 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eec36046-1daf-4f76-9a9e-3d9001830bb2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hmcxg\" (UID: \"eec36046-1daf-4f76-9a9e-3d9001830bb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.915980 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-z7h4m"] Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.920780 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6t6lv"] Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.925374 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nclc8\" (UniqueName: \"kubernetes.io/projected/71e981d5-d657-46ef-9ec0-4ddad58be6ee-kube-api-access-nclc8\") pod \"marketplace-operator-79b997595-rpftn\" (UID: \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\") " pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.927461 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr5cg\" (UniqueName: \"kubernetes.io/projected/beb8a379-4733-45b5-b31b-0e02eb9d5649-kube-api-access-sr5cg\") pod \"machine-approver-56656f9798-44428\" (UID: \"beb8a379-4733-45b5-b31b-0e02eb9d5649\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929083 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929205 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ecc96f36-7ea7-4beb-977b-ae803b665b16-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kqxb6\" (UID: \"ecc96f36-7ea7-4beb-977b-ae803b665b16\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929261 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/29fc0965-6c9b-455f-b83e-f2fd1e04c6dc-metrics-tls\") pod \"dns-default-k424z\" (UID: \"29fc0965-6c9b-455f-b83e-f2fd1e04c6dc\") " pod="openshift-dns/dns-default-k424z" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929294 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/602912b1-d12a-406a-ba6e-e0d94dc62038-config-volume\") pod \"collect-profiles-29563200-xtxcb\" (UID: \"602912b1-d12a-406a-ba6e-e0d94dc62038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929318 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a-config\") pod \"service-ca-operator-777779d784-f8svz\" (UID: \"f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929353 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jslb\" (UID: \"79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929378 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd950f1d-4d97-4ed4-8d9c-04245019bbdb-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kf9gt\" (UID: \"bd950f1d-4d97-4ed4-8d9c-04245019bbdb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929399 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-csi-data-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929419 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29fc0965-6c9b-455f-b83e-f2fd1e04c6dc-config-volume\") pod \"dns-default-k424z\" (UID: \"29fc0965-6c9b-455f-b83e-f2fd1e04c6dc\") " pod="openshift-dns/dns-default-k424z" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929440 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-plugins-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929470 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9x64\" (UniqueName: \"kubernetes.io/projected/799072d7-3ce6-4742-be4c-065ecea7e24c-kube-api-access-h9x64\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929502 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84bde919-6dad-4080-977f-81cda3119a89-config\") pod \"kube-controller-manager-operator-78b949d7b-w4hpj\" (UID: \"84bde919-6dad-4080-977f-81cda3119a89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929526 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8fvq\" (UniqueName: \"kubernetes.io/projected/29fc0965-6c9b-455f-b83e-f2fd1e04c6dc-kube-api-access-v8fvq\") pod \"dns-default-k424z\" (UID: \"29fc0965-6c9b-455f-b83e-f2fd1e04c6dc\") " pod="openshift-dns/dns-default-k424z" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929546 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-socket-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929567 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7ddf406d-94c4-4ce9-b593-aebe0fca2b51-srv-cert\") pod \"olm-operator-6b444d44fb-8qf4k\" (UID: \"7ddf406d-94c4-4ce9-b593-aebe0fca2b51\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929609 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8-tmpfs\") pod \"packageserver-d55dfcdfc-9xkhq\" (UID: \"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929631 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjvwm\" (UniqueName: \"kubernetes.io/projected/d3f900da-08ff-4983-89b7-4dea2613f0fd-kube-api-access-wjvwm\") pod \"service-ca-9c57cc56f-gx8mh\" (UID: \"d3f900da-08ff-4983-89b7-4dea2613f0fd\") " pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929666 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jslb\" (UID: \"79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929690 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ecc96f36-7ea7-4beb-977b-ae803b665b16-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kqxb6\" (UID: \"ecc96f36-7ea7-4beb-977b-ae803b665b16\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929712 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecc96f36-7ea7-4beb-977b-ae803b665b16-config\") pod \"kube-apiserver-operator-766d6c64bb-kqxb6\" (UID: \"ecc96f36-7ea7-4beb-977b-ae803b665b16\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929732 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a-serving-cert\") pod \"service-ca-operator-777779d784-f8svz\" (UID: \"f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929754 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8fjh\" (UniqueName: \"kubernetes.io/projected/bd950f1d-4d97-4ed4-8d9c-04245019bbdb-kube-api-access-t8fjh\") pod \"kube-storage-version-migrator-operator-b67b599dd-kf9gt\" (UID: \"bd950f1d-4d97-4ed4-8d9c-04245019bbdb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929780 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f9476891-3ac9-4914-9261-7d82a3367a00-certs\") pod \"machine-config-server-pch42\" (UID: \"f9476891-3ac9-4914-9261-7d82a3367a00\") " pod="openshift-machine-config-operator/machine-config-server-pch42" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929800 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84bde919-6dad-4080-977f-81cda3119a89-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-w4hpj\" (UID: \"84bde919-6dad-4080-977f-81cda3119a89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.930544 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jslb\" (UID: \"79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.930560 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84bde919-6dad-4080-977f-81cda3119a89-config\") pod \"kube-controller-manager-operator-78b949d7b-w4hpj\" (UID: \"84bde919-6dad-4080-977f-81cda3119a89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" Mar 18 00:11:45 crc kubenswrapper[4983]: E0318 00:11:45.930626 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:46.430611642 +0000 UTC m=+224.327986252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.929822 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8-apiservice-cert\") pod \"packageserver-d55dfcdfc-9xkhq\" (UID: \"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931062 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1321ddb2-5769-482f-b4dc-1971dad30fdf-profile-collector-cert\") pod \"catalog-operator-68c6474976-wsk9g\" (UID: \"1321ddb2-5769-482f-b4dc-1971dad30fdf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931103 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd950f1d-4d97-4ed4-8d9c-04245019bbdb-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kf9gt\" (UID: \"bd950f1d-4d97-4ed4-8d9c-04245019bbdb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931146 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5qrv\" (UniqueName: \"kubernetes.io/projected/63bc8988-a69a-4385-8ff3-45edafde495a-kube-api-access-p5qrv\") pod \"ingress-canary-b8gtz\" (UID: \"63bc8988-a69a-4385-8ff3-45edafde495a\") " pod="openshift-ingress-canary/ingress-canary-b8gtz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931170 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlv5c\" (UniqueName: \"kubernetes.io/projected/da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8-kube-api-access-nlv5c\") pod \"packageserver-d55dfcdfc-9xkhq\" (UID: \"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931200 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-mountpoint-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931223 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r824j\" (UniqueName: \"kubernetes.io/projected/602912b1-d12a-406a-ba6e-e0d94dc62038-kube-api-access-r824j\") pod \"collect-profiles-29563200-xtxcb\" (UID: \"602912b1-d12a-406a-ba6e-e0d94dc62038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931245 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jslb\" (UID: \"79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931269 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7237321-db39-4ed6-a22e-b3b57c5b6be6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-96hcz\" (UID: \"d7237321-db39-4ed6-a22e-b3b57c5b6be6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931296 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7ddf406d-94c4-4ce9-b593-aebe0fca2b51-profile-collector-cert\") pod \"olm-operator-6b444d44fb-8qf4k\" (UID: \"7ddf406d-94c4-4ce9-b593-aebe0fca2b51\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931339 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rf9n\" (UniqueName: \"kubernetes.io/projected/f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a-kube-api-access-6rf9n\") pod \"service-ca-operator-777779d784-f8svz\" (UID: \"f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931365 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84bde919-6dad-4080-977f-81cda3119a89-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-w4hpj\" (UID: \"84bde919-6dad-4080-977f-81cda3119a89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931387 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/602912b1-d12a-406a-ba6e-e0d94dc62038-secret-volume\") pod \"collect-profiles-29563200-xtxcb\" (UID: \"602912b1-d12a-406a-ba6e-e0d94dc62038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931414 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/63bc8988-a69a-4385-8ff3-45edafde495a-cert\") pod \"ingress-canary-b8gtz\" (UID: \"63bc8988-a69a-4385-8ff3-45edafde495a\") " pod="openshift-ingress-canary/ingress-canary-b8gtz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931429 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd950f1d-4d97-4ed4-8d9c-04245019bbdb-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kf9gt\" (UID: \"bd950f1d-4d97-4ed4-8d9c-04245019bbdb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931439 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kplv\" (UniqueName: \"kubernetes.io/projected/efeae5ef-dba6-4031-83be-8f493fe853b9-kube-api-access-5kplv\") pod \"auto-csr-approver-29563210-lptd5\" (UID: \"efeae5ef-dba6-4031-83be-8f493fe853b9\") " pod="openshift-infra/auto-csr-approver-29563210-lptd5" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931481 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-registration-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931514 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1321ddb2-5769-482f-b4dc-1971dad30fdf-srv-cert\") pod \"catalog-operator-68c6474976-wsk9g\" (UID: \"1321ddb2-5769-482f-b4dc-1971dad30fdf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931537 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cs7v\" (UniqueName: \"kubernetes.io/projected/1321ddb2-5769-482f-b4dc-1971dad30fdf-kube-api-access-9cs7v\") pod \"catalog-operator-68c6474976-wsk9g\" (UID: \"1321ddb2-5769-482f-b4dc-1971dad30fdf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931561 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtxh2\" (UniqueName: \"kubernetes.io/projected/f9476891-3ac9-4914-9261-7d82a3367a00-kube-api-access-mtxh2\") pod \"machine-config-server-pch42\" (UID: \"f9476891-3ac9-4914-9261-7d82a3367a00\") " pod="openshift-machine-config-operator/machine-config-server-pch42" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931586 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj9pm\" (UniqueName: \"kubernetes.io/projected/7ddf406d-94c4-4ce9-b593-aebe0fca2b51-kube-api-access-cj9pm\") pod \"olm-operator-6b444d44fb-8qf4k\" (UID: \"7ddf406d-94c4-4ce9-b593-aebe0fca2b51\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931606 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8-webhook-cert\") pod \"packageserver-d55dfcdfc-9xkhq\" (UID: \"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931619 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8-tmpfs\") pod \"packageserver-d55dfcdfc-9xkhq\" (UID: \"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931637 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d3f900da-08ff-4983-89b7-4dea2613f0fd-signing-cabundle\") pod \"service-ca-9c57cc56f-gx8mh\" (UID: \"d3f900da-08ff-4983-89b7-4dea2613f0fd\") " pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931659 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f9476891-3ac9-4914-9261-7d82a3367a00-node-bootstrap-token\") pod \"machine-config-server-pch42\" (UID: \"f9476891-3ac9-4914-9261-7d82a3367a00\") " pod="openshift-machine-config-operator/machine-config-server-pch42" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931681 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq2gd\" (UniqueName: \"kubernetes.io/projected/d7237321-db39-4ed6-a22e-b3b57c5b6be6-kube-api-access-xq2gd\") pod \"package-server-manager-789f6589d5-96hcz\" (UID: \"d7237321-db39-4ed6-a22e-b3b57c5b6be6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931705 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d3f900da-08ff-4983-89b7-4dea2613f0fd-signing-key\") pod \"service-ca-9c57cc56f-gx8mh\" (UID: \"d3f900da-08ff-4983-89b7-4dea2613f0fd\") " pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.931990 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-socket-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.932433 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29fc0965-6c9b-455f-b83e-f2fd1e04c6dc-config-volume\") pod \"dns-default-k424z\" (UID: \"29fc0965-6c9b-455f-b83e-f2fd1e04c6dc\") " pod="openshift-dns/dns-default-k424z" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.932518 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-csi-data-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.932698 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a-config\") pod \"service-ca-operator-777779d784-f8svz\" (UID: \"f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.935447 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/602912b1-d12a-406a-ba6e-e0d94dc62038-config-volume\") pod \"collect-profiles-29563200-xtxcb\" (UID: \"602912b1-d12a-406a-ba6e-e0d94dc62038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.936602 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-plugins-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.937389 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/29fc0965-6c9b-455f-b83e-f2fd1e04c6dc-metrics-tls\") pod \"dns-default-k424z\" (UID: \"29fc0965-6c9b-455f-b83e-f2fd1e04c6dc\") " pod="openshift-dns/dns-default-k424z" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.938489 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8-webhook-cert\") pod \"packageserver-d55dfcdfc-9xkhq\" (UID: \"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.938521 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7ddf406d-94c4-4ce9-b593-aebe0fca2b51-srv-cert\") pod \"olm-operator-6b444d44fb-8qf4k\" (UID: \"7ddf406d-94c4-4ce9-b593-aebe0fca2b51\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.938832 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-mountpoint-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.939633 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7237321-db39-4ed6-a22e-b3b57c5b6be6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-96hcz\" (UID: \"d7237321-db39-4ed6-a22e-b3b57c5b6be6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.940302 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1321ddb2-5769-482f-b4dc-1971dad30fdf-srv-cert\") pod \"catalog-operator-68c6474976-wsk9g\" (UID: \"1321ddb2-5769-482f-b4dc-1971dad30fdf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.940859 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ecc96f36-7ea7-4beb-977b-ae803b665b16-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kqxb6\" (UID: \"ecc96f36-7ea7-4beb-977b-ae803b665b16\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.942601 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/799072d7-3ce6-4742-be4c-065ecea7e24c-registration-dir\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.943941 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd950f1d-4d97-4ed4-8d9c-04245019bbdb-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kf9gt\" (UID: \"bd950f1d-4d97-4ed4-8d9c-04245019bbdb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.944223 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7ddf406d-94c4-4ce9-b593-aebe0fca2b51-profile-collector-cert\") pod \"olm-operator-6b444d44fb-8qf4k\" (UID: \"7ddf406d-94c4-4ce9-b593-aebe0fca2b51\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.944644 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jslb\" (UID: \"79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.945052 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f9476891-3ac9-4914-9261-7d82a3367a00-node-bootstrap-token\") pod \"machine-config-server-pch42\" (UID: \"f9476891-3ac9-4914-9261-7d82a3367a00\") " pod="openshift-machine-config-operator/machine-config-server-pch42" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.945744 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecc96f36-7ea7-4beb-977b-ae803b665b16-config\") pod \"kube-apiserver-operator-766d6c64bb-kqxb6\" (UID: \"ecc96f36-7ea7-4beb-977b-ae803b665b16\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.945742 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1321ddb2-5769-482f-b4dc-1971dad30fdf-profile-collector-cert\") pod \"catalog-operator-68c6474976-wsk9g\" (UID: \"1321ddb2-5769-482f-b4dc-1971dad30fdf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.946163 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d3f900da-08ff-4983-89b7-4dea2613f0fd-signing-cabundle\") pod \"service-ca-9c57cc56f-gx8mh\" (UID: \"d3f900da-08ff-4983-89b7-4dea2613f0fd\") " pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.946867 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8-apiservice-cert\") pod \"packageserver-d55dfcdfc-9xkhq\" (UID: \"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.947804 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/602912b1-d12a-406a-ba6e-e0d94dc62038-secret-volume\") pod \"collect-profiles-29563200-xtxcb\" (UID: \"602912b1-d12a-406a-ba6e-e0d94dc62038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.947849 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a-serving-cert\") pod \"service-ca-operator-777779d784-f8svz\" (UID: \"f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.948085 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f9476891-3ac9-4914-9261-7d82a3367a00-certs\") pod \"machine-config-server-pch42\" (UID: \"f9476891-3ac9-4914-9261-7d82a3367a00\") " pod="openshift-machine-config-operator/machine-config-server-pch42" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.948216 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq65z\" (UniqueName: \"kubernetes.io/projected/6d53acd4-182c-4dd4-93e1-bccac2a0a522-kube-api-access-hq65z\") pod \"multus-admission-controller-857f4d67dd-85cfx\" (UID: \"6d53acd4-182c-4dd4-93e1-bccac2a0a522\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-85cfx" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.948687 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/63bc8988-a69a-4385-8ff3-45edafde495a-cert\") pod \"ingress-canary-b8gtz\" (UID: \"63bc8988-a69a-4385-8ff3-45edafde495a\") " pod="openshift-ingress-canary/ingress-canary-b8gtz" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.949230 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84bde919-6dad-4080-977f-81cda3119a89-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-w4hpj\" (UID: \"84bde919-6dad-4080-977f-81cda3119a89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.949989 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d3f900da-08ff-4983-89b7-4dea2613f0fd-signing-key\") pod \"service-ca-9c57cc56f-gx8mh\" (UID: \"d3f900da-08ff-4983-89b7-4dea2613f0fd\") " pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.950756 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt"] Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.953205 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-w4s69"] Mar 18 00:11:45 crc kubenswrapper[4983]: W0318 00:11:45.957388 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66e8121d_9422_4b23_93e9_9f4d350c149f.slice/crio-cc53b56e437c38c8592783cbafa55e920436da4a14527406ce8715bbf6dda778 WatchSource:0}: Error finding container cc53b56e437c38c8592783cbafa55e920436da4a14527406ce8715bbf6dda778: Status 404 returned error can't find the container with id cc53b56e437c38c8592783cbafa55e920436da4a14527406ce8715bbf6dda778 Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.963458 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln6dp\" (UniqueName: \"kubernetes.io/projected/49880d65-bff4-45d1-8fb7-907e524568fc-kube-api-access-ln6dp\") pod \"machine-config-controller-84d6567774-99dq7\" (UID: \"49880d65-bff4-45d1-8fb7-907e524568fc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" Mar 18 00:11:45 crc kubenswrapper[4983]: I0318 00:11:45.986812 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvpqz\" (UniqueName: \"kubernetes.io/projected/eec36046-1daf-4f76-9a9e-3d9001830bb2-kube-api-access-hvpqz\") pod \"cluster-image-registry-operator-dc59b4c8b-hmcxg\" (UID: \"eec36046-1daf-4f76-9a9e-3d9001830bb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.011871 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.028963 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7drlm\" (UniqueName: \"kubernetes.io/projected/dafd0e48-880c-44fc-b8c5-8056a37dac27-kube-api-access-7drlm\") pod \"router-default-5444994796-ncss9\" (UID: \"dafd0e48-880c-44fc-b8c5-8056a37dac27\") " pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.034817 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.035373 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:46.535358638 +0000 UTC m=+224.432733248 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.050578 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-bound-sa-token\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.067894 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68k9x\" (UniqueName: \"kubernetes.io/projected/eb9a12c3-1119-4b1e-8d32-0a9696a00392-kube-api-access-68k9x\") pod \"downloads-7954f5f757-87bh5\" (UID: \"eb9a12c3-1119-4b1e-8d32-0a9696a00392\") " pod="openshift-console/downloads-7954f5f757-87bh5" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.083245 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4dn4m"] Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.088008 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-65l25"] Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.092841 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zlqph"] Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.093394 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c77mr\" (UniqueName: \"kubernetes.io/projected/eaebd031-fb05-489b-8be1-e4e115bdcb7d-kube-api-access-c77mr\") pod \"machine-config-operator-74547568cd-9h7fc\" (UID: \"eaebd031-fb05-489b-8be1-e4e115bdcb7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:46 crc kubenswrapper[4983]: W0318 00:11:46.097813 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fbfc372_868f_41d5_81d0_efb8a3e88023.slice/crio-27c5c46753c7de3dfecb12ffdbe1313f1f0a4c391dc855c9e48e881dc51a2516 WatchSource:0}: Error finding container 27c5c46753c7de3dfecb12ffdbe1313f1f0a4c391dc855c9e48e881dc51a2516: Status 404 returned error can't find the container with id 27c5c46753c7de3dfecb12ffdbe1313f1f0a4c391dc855c9e48e881dc51a2516 Mar 18 00:11:46 crc kubenswrapper[4983]: W0318 00:11:46.103919 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b2013fc_d64f_4f21_be62_e17f3855f6c1.slice/crio-ef176c7de73ae407113e93e01247788d708a5b67f4bfc39d062abeea19177d96 WatchSource:0}: Error finding container ef176c7de73ae407113e93e01247788d708a5b67f4bfc39d062abeea19177d96: Status 404 returned error can't find the container with id ef176c7de73ae407113e93e01247788d708a5b67f4bfc39d062abeea19177d96 Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.107677 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqrff\" (UniqueName: \"kubernetes.io/projected/a4e42b8f-7120-4b47-b8ed-725640893464-kube-api-access-lqrff\") pod \"control-plane-machine-set-operator-78cbb6b69f-q72ss\" (UID: \"a4e42b8f-7120-4b47-b8ed-725640893464\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.125506 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a57b48db-a186-4a73-b139-a9249a19a3ed-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zsklf\" (UID: \"a57b48db-a186-4a73-b139-a9249a19a3ed\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.133378 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-wxstd"] Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.133534 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-87bh5" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.140461 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.145017 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.145109 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:46.645072552 +0000 UTC m=+224.542447162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.145282 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.145806 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:46.645798174 +0000 UTC m=+224.543172784 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.149646 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6grh\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-kube-api-access-l6grh\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.160639 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.171133 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.171387 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8fvq\" (UniqueName: \"kubernetes.io/projected/29fc0965-6c9b-455f-b83e-f2fd1e04c6dc-kube-api-access-v8fvq\") pod \"dns-default-k424z\" (UID: \"29fc0965-6c9b-455f-b83e-f2fd1e04c6dc\") " pod="openshift-dns/dns-default-k424z" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.171628 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h74lp"] Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.175185 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.179077 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4"] Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.179156 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4"] Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.182848 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85"] Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.183041 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.184174 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js"] Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.185837 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjvwm\" (UniqueName: \"kubernetes.io/projected/d3f900da-08ff-4983-89b7-4dea2613f0fd-kube-api-access-wjvwm\") pod \"service-ca-9c57cc56f-gx8mh\" (UID: \"d3f900da-08ff-4983-89b7-4dea2613f0fd\") " pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.190212 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.198047 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-85cfx" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.205220 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9x64\" (UniqueName: \"kubernetes.io/projected/799072d7-3ce6-4742-be4c-065ecea7e24c-kube-api-access-h9x64\") pod \"csi-hostpathplugin-kzks7\" (UID: \"799072d7-3ce6-4742-be4c-065ecea7e24c\") " pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:46 crc kubenswrapper[4983]: W0318 00:11:46.213381 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53bc5818_4983_4f5f_8a3c_edc02aff5325.slice/crio-38d31c2d714323d93d3e1883d1ebda0e3ece2644f4165694609632d33aec4974 WatchSource:0}: Error finding container 38d31c2d714323d93d3e1883d1ebda0e3ece2644f4165694609632d33aec4974: Status 404 returned error can't find the container with id 38d31c2d714323d93d3e1883d1ebda0e3ece2644f4165694609632d33aec4974 Mar 18 00:11:46 crc kubenswrapper[4983]: W0318 00:11:46.213858 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79f09f2f_fe8b_4340_a1fc_c16f212b7e31.slice/crio-f9be1dfdf4538c4f9896810223c0ce0a022f1129c9d2b832403abfec9c71d86b WatchSource:0}: Error finding container f9be1dfdf4538c4f9896810223c0ce0a022f1129c9d2b832403abfec9c71d86b: Status 404 returned error can't find the container with id f9be1dfdf4538c4f9896810223c0ce0a022f1129c9d2b832403abfec9c71d86b Mar 18 00:11:46 crc kubenswrapper[4983]: W0318 00:11:46.214099 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0774a9cb_d9cb_45a2_89e1_384902cf8f94.slice/crio-9a095257a38fc8517d53f385bbbfba4c70e41aa39cf57e38071dcb3cb254bc1a WatchSource:0}: Error finding container 9a095257a38fc8517d53f385bbbfba4c70e41aa39cf57e38071dcb3cb254bc1a: Status 404 returned error can't find the container with id 9a095257a38fc8517d53f385bbbfba4c70e41aa39cf57e38071dcb3cb254bc1a Mar 18 00:11:46 crc kubenswrapper[4983]: W0318 00:11:46.217970 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c4f3aae_cfe4_4935_b1d6_3c7be1557e5d.slice/crio-94a5ec1ff11e094545a5a0e4cb15ca88ee65b0d0ec1cb036c60c193b827e180a WatchSource:0}: Error finding container 94a5ec1ff11e094545a5a0e4cb15ca88ee65b0d0ec1cb036c60c193b827e180a: Status 404 returned error can't find the container with id 94a5ec1ff11e094545a5a0e4cb15ca88ee65b0d0ec1cb036c60c193b827e180a Mar 18 00:11:46 crc kubenswrapper[4983]: W0318 00:11:46.221532 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c447db7_c45e_4e01_925a_6ac56ec2aee2.slice/crio-183d78e94ac0c52be1b7af190067fe2a4d23b2da0d72607ae16fdde53800dc6f WatchSource:0}: Error finding container 183d78e94ac0c52be1b7af190067fe2a4d23b2da0d72607ae16fdde53800dc6f: Status 404 returned error can't find the container with id 183d78e94ac0c52be1b7af190067fe2a4d23b2da0d72607ae16fdde53800dc6f Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.226403 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.227545 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlv5c\" (UniqueName: \"kubernetes.io/projected/da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8-kube-api-access-nlv5c\") pod \"packageserver-d55dfcdfc-9xkhq\" (UID: \"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.232687 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wxstd" event={"ID":"b876a42a-856b-4530-a601-3ca35db1078a","Type":"ContainerStarted","Data":"b98bc2a17c1be4979ddd5c78094cbcee878eb115baa546ffd79196a192a757ff"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.234980 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29563200-6l8ds" event={"ID":"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8","Type":"ContainerStarted","Data":"b9dfc5a4c8fa5a62b5be3aa151ae19a10c0341e65307da84c928ebcfd7457397"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.235055 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29563200-6l8ds" event={"ID":"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8","Type":"ContainerStarted","Data":"3fce4956773ff2e831fd8c56997a98538f386e09f45740556fa4487b7ea44a83"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.246893 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.247050 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:46.747018564 +0000 UTC m=+224.644393174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.247288 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.247710 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:46.747700735 +0000 UTC m=+224.645075345 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.248198 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cs7v\" (UniqueName: \"kubernetes.io/projected/1321ddb2-5769-482f-b4dc-1971dad30fdf-kube-api-access-9cs7v\") pod \"catalog-operator-68c6474976-wsk9g\" (UID: \"1321ddb2-5769-482f-b4dc-1971dad30fdf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.250711 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" event={"ID":"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d","Type":"ContainerStarted","Data":"94a5ec1ff11e094545a5a0e4cb15ca88ee65b0d0ec1cb036c60c193b827e180a"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.257696 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-65l25" event={"ID":"6375157a-012b-441e-b97b-39d6830cef68","Type":"ContainerStarted","Data":"05295044bdc4e4783f11e2362f63dec0f06d271b74df79d962d063ec595eb869"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.271035 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtxh2\" (UniqueName: \"kubernetes.io/projected/f9476891-3ac9-4914-9261-7d82a3367a00-kube-api-access-mtxh2\") pod \"machine-config-server-pch42\" (UID: \"f9476891-3ac9-4914-9261-7d82a3367a00\") " pod="openshift-machine-config-operator/machine-config-server-pch42" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.274277 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" event={"ID":"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8","Type":"ContainerStarted","Data":"a2544a60200e6e0dfebfa517f0865d521965c34b12d9bba1972904432e3e4969"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.274323 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" event={"ID":"e0f5e525-af83-4ee9-96c6-a0ec92c5e1b8","Type":"ContainerStarted","Data":"dc8c9a9aadd9d5128bef45a33a456dd5abae0fbcd88dbe6a93306010664b7a6b"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.280278 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" event={"ID":"4fbfc372-868f-41d5-81d0-efb8a3e88023","Type":"ContainerStarted","Data":"27c5c46753c7de3dfecb12ffdbe1313f1f0a4c391dc855c9e48e881dc51a2516"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.283435 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" event={"ID":"0b2013fc-d64f-4f21-be62-e17f3855f6c1","Type":"ContainerStarted","Data":"ef176c7de73ae407113e93e01247788d708a5b67f4bfc39d062abeea19177d96"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.284302 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-w4s69" event={"ID":"6e69d708-f2df-436e-a5ec-41848aecbace","Type":"ContainerStarted","Data":"1614e749b6d06c9963f1aa101484210a9f1fea7dbe88ec14001824417d684d3c"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.285107 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" event={"ID":"0774a9cb-d9cb-45a2-89e1-384902cf8f94","Type":"ContainerStarted","Data":"9a095257a38fc8517d53f385bbbfba4c70e41aa39cf57e38071dcb3cb254bc1a"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.289828 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" event={"ID":"5c447db7-c45e-4e01-925a-6ac56ec2aee2","Type":"ContainerStarted","Data":"183d78e94ac0c52be1b7af190067fe2a4d23b2da0d72607ae16fdde53800dc6f"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.294386 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h74lp" event={"ID":"53bc5818-4983-4f5f-8a3c-edc02aff5325","Type":"ContainerStarted","Data":"38d31c2d714323d93d3e1883d1ebda0e3ece2644f4165694609632d33aec4974"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.297626 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-gvd97"] Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.298066 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj9pm\" (UniqueName: \"kubernetes.io/projected/7ddf406d-94c4-4ce9-b593-aebe0fca2b51-kube-api-access-cj9pm\") pod \"olm-operator-6b444d44fb-8qf4k\" (UID: \"7ddf406d-94c4-4ce9-b593-aebe0fca2b51\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.298904 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj"] Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.302260 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-slqc5"] Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.304275 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" event={"ID":"beb8a379-4733-45b5-b31b-0e02eb9d5649","Type":"ContainerStarted","Data":"beddc79ab5395af9f46a35f8299ad57be4795b61a6b91499c1af1c40fd3de197"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.305458 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" event={"ID":"66e8121d-9422-4b23-93e9-9f4d350c149f","Type":"ContainerStarted","Data":"cc53b56e437c38c8592783cbafa55e920436da4a14527406ce8715bbf6dda778"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.306672 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" event={"ID":"744c22bb-2f35-494b-b352-ff71691bd0af","Type":"ContainerStarted","Data":"e392bf937d16711f20a240fc951127dce585f9bf75831fde980b9447f25b40ad"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.306688 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" event={"ID":"744c22bb-2f35-494b-b352-ff71691bd0af","Type":"ContainerStarted","Data":"5c6d57161127ee14e7cf4c213041295a3273ed0c5e16022c2b9435b72539b391"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.308705 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" event={"ID":"79f09f2f-fe8b-4340-a1fc-c16f212b7e31","Type":"ContainerStarted","Data":"f9be1dfdf4538c4f9896810223c0ce0a022f1129c9d2b832403abfec9c71d86b"} Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.317547 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ecc96f36-7ea7-4beb-977b-ae803b665b16-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kqxb6\" (UID: \"ecc96f36-7ea7-4beb-977b-ae803b665b16\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.317561 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-pch42" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.322702 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r824j\" (UniqueName: \"kubernetes.io/projected/602912b1-d12a-406a-ba6e-e0d94dc62038-kube-api-access-r824j\") pod \"collect-profiles-29563200-xtxcb\" (UID: \"602912b1-d12a-406a-ba6e-e0d94dc62038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.322988 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-k424z" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.329983 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-kzks7" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.351703 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kplv\" (UniqueName: \"kubernetes.io/projected/efeae5ef-dba6-4031-83be-8f493fe853b9-kube-api-access-5kplv\") pod \"auto-csr-approver-29563210-lptd5\" (UID: \"efeae5ef-dba6-4031-83be-8f493fe853b9\") " pod="openshift-infra/auto-csr-approver-29563210-lptd5" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.352340 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.353300 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:46.853285715 +0000 UTC m=+224.750660325 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.354802 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:46 crc kubenswrapper[4983]: W0318 00:11:46.377360 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04baa777_3a33_4b5a_9808_99a84b1b8005.slice/crio-3e6beffae814819f4769ec65c3cedca70ffe31206925285054972d7b8b43f43c WatchSource:0}: Error finding container 3e6beffae814819f4769ec65c3cedca70ffe31206925285054972d7b8b43f43c: Status 404 returned error can't find the container with id 3e6beffae814819f4769ec65c3cedca70ffe31206925285054972d7b8b43f43c Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.384601 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5qrv\" (UniqueName: \"kubernetes.io/projected/63bc8988-a69a-4385-8ff3-45edafde495a-kube-api-access-p5qrv\") pod \"ingress-canary-b8gtz\" (UID: \"63bc8988-a69a-4385-8ff3-45edafde495a\") " pod="openshift-ingress-canary/ingress-canary-b8gtz" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.394239 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jslb\" (UID: \"79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.409802 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq2gd\" (UniqueName: \"kubernetes.io/projected/d7237321-db39-4ed6-a22e-b3b57c5b6be6-kube-api-access-xq2gd\") pod \"package-server-manager-789f6589d5-96hcz\" (UID: \"d7237321-db39-4ed6-a22e-b3b57c5b6be6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.426035 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.429052 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rf9n\" (UniqueName: \"kubernetes.io/projected/f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a-kube-api-access-6rf9n\") pod \"service-ca-operator-777779d784-f8svz\" (UID: \"f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.443980 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-87bh5"] Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.446391 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8fjh\" (UniqueName: \"kubernetes.io/projected/bd950f1d-4d97-4ed4-8d9c-04245019bbdb-kube-api-access-t8fjh\") pod \"kube-storage-version-migrator-operator-b67b599dd-kf9gt\" (UID: \"bd950f1d-4d97-4ed4-8d9c-04245019bbdb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.453713 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.454345 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:46.954323421 +0000 UTC m=+224.851698031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.468805 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84bde919-6dad-4080-977f-81cda3119a89-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-w4hpj\" (UID: \"84bde919-6dad-4080-977f-81cda3119a89\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.505221 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.511973 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.520624 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.534669 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.551227 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563210-lptd5" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.555274 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.555582 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.05551623 +0000 UTC m=+224.952890850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.555697 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.556361 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.056328615 +0000 UTC m=+224.953703395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.558625 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.576333 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.582706 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.590650 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.598246 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b8gtz" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.606732 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" Mar 18 00:11:46 crc kubenswrapper[4983]: W0318 00:11:46.616275 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb9a12c3_1119_4b1e_8d32_0a9696a00392.slice/crio-d979d9adfaef643118d21a0d7bbf23657b412223cae99eafea8c358f39cbe8e4 WatchSource:0}: Error finding container d979d9adfaef643118d21a0d7bbf23657b412223cae99eafea8c358f39cbe8e4: Status 404 returned error can't find the container with id d979d9adfaef643118d21a0d7bbf23657b412223cae99eafea8c358f39cbe8e4 Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.657044 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.657675 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.157648918 +0000 UTC m=+225.055023528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.658474 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.658919 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.158902676 +0000 UTC m=+225.056277286 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.760126 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.760767 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.260730754 +0000 UTC m=+225.158105384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.760822 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.761237 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.261227249 +0000 UTC m=+225.158601859 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.862095 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.862272 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.362242224 +0000 UTC m=+225.259616834 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.862578 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.863052 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.363039008 +0000 UTC m=+225.260413618 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.964506 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.965075 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.465062852 +0000 UTC m=+225.362437462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.965415 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:46 crc kubenswrapper[4983]: I0318 00:11:46.964915 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-z7h4m" podStartSLOduration=162.964889977 podStartE2EDuration="2m42.964889977s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:46.952699583 +0000 UTC m=+224.850074213" watchObservedRunningTime="2026-03-18 00:11:46.964889977 +0000 UTC m=+224.862264587" Mar 18 00:11:46 crc kubenswrapper[4983]: E0318 00:11:46.965847 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.465835535 +0000 UTC m=+225.363210145 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.013701 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.041818 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-k424z"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.068523 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:47 crc kubenswrapper[4983]: E0318 00:11:47.068805 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.568771997 +0000 UTC m=+225.466146607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.069212 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:47 crc kubenswrapper[4983]: E0318 00:11:47.069708 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.569699805 +0000 UTC m=+225.467074415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.128451 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gx8mh"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.131125 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rpftn"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.143831 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.150028 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-kzks7"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.150486 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.171805 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:47 crc kubenswrapper[4983]: E0318 00:11:47.172194 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.672177153 +0000 UTC m=+225.569551753 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.182692 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-85cfx"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.186652 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.207941 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.218606 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.220038 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.273669 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:47 crc kubenswrapper[4983]: E0318 00:11:47.274740 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.774682092 +0000 UTC m=+225.672056702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.318510 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" event={"ID":"71e981d5-d657-46ef-9ec0-4ddad58be6ee","Type":"ContainerStarted","Data":"35f502854b30ecc086620c65928c61fcbce8a272d8f90335f219d16f976ab38c"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.322299 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" event={"ID":"651bef83-00db-40d1-b096-247b3c645859","Type":"ContainerStarted","Data":"21f6567fbbe38e46d438c001f65dc260552a57db6fb851c467440c0c17c292c3"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.342763 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.343388 4983 generic.go:334] "Generic (PLEG): container finished" podID="744c22bb-2f35-494b-b352-ff71691bd0af" containerID="e392bf937d16711f20a240fc951127dce585f9bf75831fde980b9447f25b40ad" exitCode=0 Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.343475 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" event={"ID":"744c22bb-2f35-494b-b352-ff71691bd0af","Type":"ContainerDied","Data":"e392bf937d16711f20a240fc951127dce585f9bf75831fde980b9447f25b40ad"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.362595 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" event={"ID":"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d","Type":"ContainerStarted","Data":"d9eb68e8bcc8979fda5ff8f269ddbb7bc410310bba6eba0fe60bd4b384c28556"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.363138 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.371191 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" event={"ID":"beb8a379-4733-45b5-b31b-0e02eb9d5649","Type":"ContainerStarted","Data":"d9f62badc38015a817813ef6b8326a19ddf6583fae7aa99593305a43e653888a"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.374133 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:47 crc kubenswrapper[4983]: E0318 00:11:47.374453 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.874436739 +0000 UTC m=+225.771811349 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.379250 4983 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-5j7r4 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.379303 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" podUID="5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.379263 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-65l25" event={"ID":"6375157a-012b-441e-b97b-39d6830cef68","Type":"ContainerStarted","Data":"a889016ad154a180f550264d00b88be58cf1b17c430b3e51d17966e21b5debf3"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.388778 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" event={"ID":"eaebd031-fb05-489b-8be1-e4e115bdcb7d","Type":"ContainerStarted","Data":"b8f608a08eb94d19e02aadad5c9ffabc37584a6343d4838dec3cd4418005d362"} Mar 18 00:11:47 crc kubenswrapper[4983]: W0318 00:11:47.404289 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4e42b8f_7120_4b47_b8ed_725640893464.slice/crio-d03c09ca04cf050994a33aeba6c1116b0b30a04c97329ed220a04d695a66b575 WatchSource:0}: Error finding container d03c09ca04cf050994a33aeba6c1116b0b30a04c97329ed220a04d695a66b575: Status 404 returned error can't find the container with id d03c09ca04cf050994a33aeba6c1116b0b30a04c97329ed220a04d695a66b575 Mar 18 00:11:47 crc kubenswrapper[4983]: W0318 00:11:47.406107 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d53acd4_182c_4dd4_93e1_bccac2a0a522.slice/crio-2442cff179592047d94ee84e0fcaf4adea138a9f79d1a42ee2e9aca2e7207404 WatchSource:0}: Error finding container 2442cff179592047d94ee84e0fcaf4adea138a9f79d1a42ee2e9aca2e7207404: Status 404 returned error can't find the container with id 2442cff179592047d94ee84e0fcaf4adea138a9f79d1a42ee2e9aca2e7207404 Mar 18 00:11:47 crc kubenswrapper[4983]: W0318 00:11:47.408177 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49880d65_bff4_45d1_8fb7_907e524568fc.slice/crio-96f34bbf6195db9839b1035a47697e19f2d2f17a4fcdc23bdc141eaa772db0d3 WatchSource:0}: Error finding container 96f34bbf6195db9839b1035a47697e19f2d2f17a4fcdc23bdc141eaa772db0d3: Status 404 returned error can't find the container with id 96f34bbf6195db9839b1035a47697e19f2d2f17a4fcdc23bdc141eaa772db0d3 Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.413164 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-87bh5" event={"ID":"eb9a12c3-1119-4b1e-8d32-0a9696a00392","Type":"ContainerStarted","Data":"d979d9adfaef643118d21a0d7bbf23657b412223cae99eafea8c358f39cbe8e4"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.426655 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-k424z" event={"ID":"29fc0965-6c9b-455f-b83e-f2fd1e04c6dc","Type":"ContainerStarted","Data":"bd7a82980e5fbd4be6275b0fcb098f8927b28ae5568144cb3a55e92119826cc3"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.429421 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-pch42" event={"ID":"f9476891-3ac9-4914-9261-7d82a3367a00","Type":"ContainerStarted","Data":"18c3c652c94c427bd7d9e8da409f259a5998acd4296256d8eb1aa8c1dab99085"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.430719 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj" event={"ID":"812c885e-98f8-47cb-afd6-5fb537bea3e3","Type":"ContainerStarted","Data":"9505eff09668009ef0dc722b367c6887c6677076b6493d967a0f7e1836cf62c8"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.438796 4983 generic.go:334] "Generic (PLEG): container finished" podID="53bc5818-4983-4f5f-8a3c-edc02aff5325" containerID="8a23bcbd90d23b2ee780be829785e1054d0ed6d0b507134c26201207f88af021" exitCode=0 Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.438901 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h74lp" event={"ID":"53bc5818-4983-4f5f-8a3c-edc02aff5325","Type":"ContainerDied","Data":"8a23bcbd90d23b2ee780be829785e1054d0ed6d0b507134c26201207f88af021"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.445521 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" event={"ID":"0b2013fc-d64f-4f21-be62-e17f3855f6c1","Type":"ContainerStarted","Data":"3cb762f5e037bbf892d7687651e70656cb61ca252d75cac1cc7887be5bad691b"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.446834 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.455164 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" event={"ID":"d3f900da-08ff-4983-89b7-4dea2613f0fd","Type":"ContainerStarted","Data":"fb7eaaae551e57bf69e5d8a6f246c744a1e422288ac96050b1e31927566a06fa"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.458084 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kzks7" event={"ID":"799072d7-3ce6-4742-be4c-065ecea7e24c","Type":"ContainerStarted","Data":"ff6b1d45bcfe51676f0e50de7977bb841492abda793b2bbf4df502bacd64ec80"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.463696 4983 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-zlqph container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.463737 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" podUID="0b2013fc-d64f-4f21-be62-e17f3855f6c1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.470973 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-w4s69" event={"ID":"6e69d708-f2df-436e-a5ec-41848aecbace","Type":"ContainerStarted","Data":"d48ed39e9f16223f9729cbbbb46fd74260d7bad888af591a53e2ab9a8b3b3fd8"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.475681 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:47 crc kubenswrapper[4983]: E0318 00:11:47.475987 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:47.975974969 +0000 UTC m=+225.873349579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.525724 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ncss9" event={"ID":"dafd0e48-880c-44fc-b8c5-8056a37dac27","Type":"ContainerStarted","Data":"6d99f2f9854c129f94d37450e9408a8a9a4adc5d4b9953f0e7ee741e77716441"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.525766 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ncss9" event={"ID":"dafd0e48-880c-44fc-b8c5-8056a37dac27","Type":"ContainerStarted","Data":"0005b272f7c49359adbbd73e5cda62e44763b126512b9cf697f5d071ddea0d28"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.542808 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" event={"ID":"0774a9cb-d9cb-45a2-89e1-384902cf8f94","Type":"ContainerStarted","Data":"96e054ab20b2a20022435e1fa3b784ceac9a696a7bb579a935a4f9a02a7a2f90"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.545183 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" event={"ID":"4fbfc372-868f-41d5-81d0-efb8a3e88023","Type":"ContainerStarted","Data":"d596dfa769b2ea64f1900d6e6139e7e729b927b5fa32360c3cf245a914d76db5"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.545224 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" event={"ID":"4fbfc372-868f-41d5-81d0-efb8a3e88023","Type":"ContainerStarted","Data":"a204dabd3a8b223323deaa87ee861344ee5e612067df59c748f8ebb59a5df526"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.566293 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" event={"ID":"5c447db7-c45e-4e01-925a-6ac56ec2aee2","Type":"ContainerStarted","Data":"490801ea7bacf5bcdfd375d5743136c2158843eb85d5ee41432d9f67ff75137c"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.576462 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:47 crc kubenswrapper[4983]: E0318 00:11:47.577470 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:48.077454157 +0000 UTC m=+225.974828767 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.586445 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" event={"ID":"79f09f2f-fe8b-4340-a1fc-c16f212b7e31","Type":"ContainerStarted","Data":"e8e6e2d79d1f43d1692006af4d7ae8f7b790b0ad96ac7cd14491111028210f07"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.598430 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" event={"ID":"66e8121d-9422-4b23-93e9-9f4d350c149f","Type":"ContainerStarted","Data":"8b0c8721863a3abbba8bdda49afb6a1e93d345e282e9df701840572628a7871d"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.598799 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.601984 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" event={"ID":"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8","Type":"ContainerStarted","Data":"7fa56c947bc1712846307adec4d35c06b6e3d22973efbc8f00ea0835726323e5"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.604762 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" event={"ID":"eec36046-1daf-4f76-9a9e-3d9001830bb2","Type":"ContainerStarted","Data":"a086761128a309e7bebbc2b1326848bcfd25cb9e1e01a973b6f01f2291c16be3"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.616850 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-slqc5" event={"ID":"04baa777-3a33-4b5a-9808-99a84b1b8005","Type":"ContainerStarted","Data":"c2092d95d1754651a362f2bf12a259268f7e8d990a23390fa2f38fc6e9a5e98b"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.617690 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-slqc5" event={"ID":"04baa777-3a33-4b5a-9808-99a84b1b8005","Type":"ContainerStarted","Data":"3e6beffae814819f4769ec65c3cedca70ffe31206925285054972d7b8b43f43c"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.619725 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.622766 4983 patch_prober.go:28] interesting pod/console-operator-58897d9998-slqc5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.622817 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-slqc5" podUID="04baa777-3a33-4b5a-9808-99a84b1b8005" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.633405 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wxstd" event={"ID":"b876a42a-856b-4530-a601-3ca35db1078a","Type":"ContainerStarted","Data":"1ca08dfb050c08a75efb86030b9192bd3b40d9dee9e36b56e44eba55f2efe4e3"} Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.679478 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:47 crc kubenswrapper[4983]: E0318 00:11:47.694101 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:48.194065717 +0000 UTC m=+226.091440327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.782091 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.784147 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k"] Mar 18 00:11:47 crc kubenswrapper[4983]: E0318 00:11:47.784727 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:48.284697872 +0000 UTC m=+226.182072482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.784865 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:47 crc kubenswrapper[4983]: E0318 00:11:47.785259 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:48.285247738 +0000 UTC m=+226.182622348 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.804264 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb"] Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.885580 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:47 crc kubenswrapper[4983]: E0318 00:11:47.886877 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:48.386859191 +0000 UTC m=+226.284233801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.895888 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29563200-6l8ds" podStartSLOduration=163.895868129 podStartE2EDuration="2m43.895868129s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:47.895463057 +0000 UTC m=+225.792837667" watchObservedRunningTime="2026-03-18 00:11:47.895868129 +0000 UTC m=+225.793242739" Mar 18 00:11:47 crc kubenswrapper[4983]: I0318 00:11:47.992023 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:47 crc kubenswrapper[4983]: E0318 00:11:47.992334 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:48.492321168 +0000 UTC m=+226.389695778 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:48 crc kubenswrapper[4983]: W0318 00:11:48.023436 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod602912b1_d12a_406a_ba6e_e0d94dc62038.slice/crio-04802dd1029d53bf5e83029b4a1b711560104f8a7f71db882592357af4d18264 WatchSource:0}: Error finding container 04802dd1029d53bf5e83029b4a1b711560104f8a7f71db882592357af4d18264: Status 404 returned error can't find the container with id 04802dd1029d53bf5e83029b4a1b711560104f8a7f71db882592357af4d18264 Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.080558 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563210-lptd5"] Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.092762 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:48 crc kubenswrapper[4983]: E0318 00:11:48.093136 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:48.593113506 +0000 UTC m=+226.490488116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.111613 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz"] Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.117492 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt"] Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.192639 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.193854 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:48 crc kubenswrapper[4983]: E0318 00:11:48.197644 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:48.697625885 +0000 UTC m=+226.595000485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.200276 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l46js" podStartSLOduration=164.200256933 podStartE2EDuration="2m44.200256933s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.153297902 +0000 UTC m=+226.050672512" watchObservedRunningTime="2026-03-18 00:11:48.200256933 +0000 UTC m=+226.097631543" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.201138 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj"] Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.203016 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b8gtz"] Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.212180 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g"] Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.222545 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:11:48 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:11:48 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:11:48 crc kubenswrapper[4983]: healthz check failed Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.222614 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.222692 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.223288 4983 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.238907 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-f8svz"] Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.241631 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-slqc5" podStartSLOduration=164.241613387 podStartE2EDuration="2m44.241613387s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.236520975 +0000 UTC m=+226.133895605" watchObservedRunningTime="2026-03-18 00:11:48.241613387 +0000 UTC m=+226.138987997" Mar 18 00:11:48 crc kubenswrapper[4983]: W0318 00:11:48.285751 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1321ddb2_5769_482f_b4dc_1971dad30fdf.slice/crio-689b7e591fd6dced9ce7ecc04d44009ec080da8e1855a84313f874b36cc3e804 WatchSource:0}: Error finding container 689b7e591fd6dced9ce7ecc04d44009ec080da8e1855a84313f874b36cc3e804: Status 404 returned error can't find the container with id 689b7e591fd6dced9ce7ecc04d44009ec080da8e1855a84313f874b36cc3e804 Mar 18 00:11:48 crc kubenswrapper[4983]: W0318 00:11:48.288470 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84bde919_6dad_4080_977f_81cda3119a89.slice/crio-8a8382e27fa475a9c2b9bd983478414d2b152ef8df2c0fdc0858b0169d92cf20 WatchSource:0}: Error finding container 8a8382e27fa475a9c2b9bd983478414d2b152ef8df2c0fdc0858b0169d92cf20: Status 404 returned error can't find the container with id 8a8382e27fa475a9c2b9bd983478414d2b152ef8df2c0fdc0858b0169d92cf20 Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.296279 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:48 crc kubenswrapper[4983]: E0318 00:11:48.296696 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:48.796677001 +0000 UTC m=+226.694051611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.313792 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" podStartSLOduration=163.313775631 podStartE2EDuration="2m43.313775631s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.310134232 +0000 UTC m=+226.207508842" watchObservedRunningTime="2026-03-18 00:11:48.313775631 +0000 UTC m=+226.211150241" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.366621 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" podStartSLOduration=164.366604097 podStartE2EDuration="2m44.366604097s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.362089323 +0000 UTC m=+226.259463953" watchObservedRunningTime="2026-03-18 00:11:48.366604097 +0000 UTC m=+226.263978707" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.399643 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:48 crc kubenswrapper[4983]: E0318 00:11:48.400073 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:48.900061166 +0000 UTC m=+226.797435776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.400860 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2xj85" podStartSLOduration=164.400838709 podStartE2EDuration="2m44.400838709s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.390800279 +0000 UTC m=+226.288174889" watchObservedRunningTime="2026-03-18 00:11:48.400838709 +0000 UTC m=+226.298213319" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.433049 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-ncss9" podStartSLOduration=163.43303096 podStartE2EDuration="2m43.43303096s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.429331309 +0000 UTC m=+226.326705919" watchObservedRunningTime="2026-03-18 00:11:48.43303096 +0000 UTC m=+226.330405570" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.485140 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" podStartSLOduration=164.485119674 podStartE2EDuration="2m44.485119674s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.481886888 +0000 UTC m=+226.379261518" watchObservedRunningTime="2026-03-18 00:11:48.485119674 +0000 UTC m=+226.382494284" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.501515 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:48 crc kubenswrapper[4983]: E0318 00:11:48.501873 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:49.001856534 +0000 UTC m=+226.899231144 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.511135 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-65l25" podStartSLOduration=164.51112314 podStartE2EDuration="2m44.51112314s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.50911333 +0000 UTC m=+226.406487940" watchObservedRunningTime="2026-03-18 00:11:48.51112314 +0000 UTC m=+226.408497750" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.610020 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:48 crc kubenswrapper[4983]: E0318 00:11:48.610305 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:49.11029291 +0000 UTC m=+227.007667520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.614475 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-4dn4m" podStartSLOduration=163.614450764 podStartE2EDuration="2m43.614450764s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.548760833 +0000 UTC m=+226.446135453" watchObservedRunningTime="2026-03-18 00:11:48.614450764 +0000 UTC m=+226.511825394" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.693255 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-pch42" event={"ID":"f9476891-3ac9-4914-9261-7d82a3367a00","Type":"ContainerStarted","Data":"d7dac84e3888977977a2f7e8433da6783982f8d208b327e0389e052d8a4c4f8d"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.714302 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b8gtz" event={"ID":"63bc8988-a69a-4385-8ff3-45edafde495a","Type":"ContainerStarted","Data":"e262e6882f332b0eec5961f7bdb1d5347d2ac6aca1517b30a076b46b444e2d29"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.714911 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:48 crc kubenswrapper[4983]: E0318 00:11:48.716004 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:49.215984803 +0000 UTC m=+227.113359413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.725830 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-pch42" podStartSLOduration=5.725813556 podStartE2EDuration="5.725813556s" podCreationTimestamp="2026-03-18 00:11:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.72459532 +0000 UTC m=+226.621969930" watchObservedRunningTime="2026-03-18 00:11:48.725813556 +0000 UTC m=+226.623188166" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.764045 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-w4s69" event={"ID":"6e69d708-f2df-436e-a5ec-41848aecbace","Type":"ContainerStarted","Data":"39477e283ecf99dbb8e07d7bb5be3261ff09714127c4bfff8640a4b91dadcc15"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.789824 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" event={"ID":"beb8a379-4733-45b5-b31b-0e02eb9d5649","Type":"ContainerStarted","Data":"c77f792a9675a18f4263522223b3e98e76441600bcf54d9e5f33236c8ac23d4a"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.791018 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-w4s69" podStartSLOduration=164.791005951 podStartE2EDuration="2m44.791005951s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.789388283 +0000 UTC m=+226.686762903" watchObservedRunningTime="2026-03-18 00:11:48.791005951 +0000 UTC m=+226.688380561" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.806360 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" event={"ID":"d7237321-db39-4ed6-a22e-b3b57c5b6be6","Type":"ContainerStarted","Data":"ad5622e807a78ccf58719d9a3aa5058d907d4aa923f65ebf22b4111b60bb39a3"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.816965 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:48 crc kubenswrapper[4983]: E0318 00:11:48.817283 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:49.317271455 +0000 UTC m=+227.214646065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.819944 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563210-lptd5" event={"ID":"efeae5ef-dba6-4031-83be-8f493fe853b9","Type":"ContainerStarted","Data":"48c5b64768b2d14183c7a618f61e947bc1f88fa07119bdef7b88bd1fc4adb3ed"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.829453 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-44428" podStartSLOduration=164.829429688 podStartE2EDuration="2m44.829429688s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.82110771 +0000 UTC m=+226.718482330" watchObservedRunningTime="2026-03-18 00:11:48.829429688 +0000 UTC m=+226.726804298" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.834352 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" event={"ID":"79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e","Type":"ContainerStarted","Data":"9016f2ff8c29327eb0e148f62a1d618f7004cee1d0067967556353739d88e819"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.834380 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" event={"ID":"79026bed-cf2a-46bb-ba6d-5bc24f1bbd1e","Type":"ContainerStarted","Data":"ad3014402f56d52a0d7f7f62e6e371b0917c142bf7b61e0f7689da22bfce0490"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.871444 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jslb" podStartSLOduration=163.871424911 podStartE2EDuration="2m43.871424911s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.870097372 +0000 UTC m=+226.767471982" watchObservedRunningTime="2026-03-18 00:11:48.871424911 +0000 UTC m=+226.768799521" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.873150 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" event={"ID":"a57b48db-a186-4a73-b139-a9249a19a3ed","Type":"ContainerStarted","Data":"705969e02ce15f963a772a809c34eab66caa545bc4cd778675638671a975a0b7"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.873180 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" event={"ID":"a57b48db-a186-4a73-b139-a9249a19a3ed","Type":"ContainerStarted","Data":"4ba2c1db1901335c15a9c9bfad2a90aa65238252c54a998e09f12fbaf78a7fc3"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.873189 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-85cfx" event={"ID":"6d53acd4-182c-4dd4-93e1-bccac2a0a522","Type":"ContainerStarted","Data":"2442cff179592047d94ee84e0fcaf4adea138a9f79d1a42ee2e9aca2e7207404"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.873305 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" event={"ID":"bd950f1d-4d97-4ed4-8d9c-04245019bbdb","Type":"ContainerStarted","Data":"54cadb95fe44a5df95c4d830de75d02ffa4fdc675e83db4b0d16bcf0cc79e021"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.918998 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:48 crc kubenswrapper[4983]: E0318 00:11:48.919214 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:49.419178986 +0000 UTC m=+227.316553596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.919305 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:48 crc kubenswrapper[4983]: E0318 00:11:48.921146 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:49.421130055 +0000 UTC m=+227.318504665 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.936600 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" event={"ID":"651bef83-00db-40d1-b096-247b3c645859","Type":"ContainerStarted","Data":"3ac6a75713c1fb53bbacb5713685ff16a359bff55c56490d21e938dbb1014a44"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.966725 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" podStartSLOduration=163.966706305 podStartE2EDuration="2m43.966706305s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.954672196 +0000 UTC m=+226.852046806" watchObservedRunningTime="2026-03-18 00:11:48.966706305 +0000 UTC m=+226.864080915" Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.989897 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss" event={"ID":"a4e42b8f-7120-4b47-b8ed-725640893464","Type":"ContainerStarted","Data":"faac03ebed97cffb0b7681a227b9e2abf116c37bb011eaee1ceabf10dc1e20f8"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.989994 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss" event={"ID":"a4e42b8f-7120-4b47-b8ed-725640893464","Type":"ContainerStarted","Data":"d03c09ca04cf050994a33aeba6c1116b0b30a04c97329ed220a04d695a66b575"} Mar 18 00:11:48 crc kubenswrapper[4983]: I0318 00:11:48.997171 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-gvd97" podStartSLOduration=164.997154493 podStartE2EDuration="2m44.997154493s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:48.993985749 +0000 UTC m=+226.891360369" watchObservedRunningTime="2026-03-18 00:11:48.997154493 +0000 UTC m=+226.894529103" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.012096 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" event={"ID":"602912b1-d12a-406a-ba6e-e0d94dc62038","Type":"ContainerStarted","Data":"04802dd1029d53bf5e83029b4a1b711560104f8a7f71db882592357af4d18264"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.023628 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:49 crc kubenswrapper[4983]: E0318 00:11:49.024821 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:49.524803738 +0000 UTC m=+227.422178348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.026656 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" event={"ID":"49880d65-bff4-45d1-8fb7-907e524568fc","Type":"ContainerStarted","Data":"db88e4eeaabab6d24cb177dc669acff073579c300b3764d778ca91fe49bb69c1"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.026708 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" event={"ID":"49880d65-bff4-45d1-8fb7-907e524568fc","Type":"ContainerStarted","Data":"96f34bbf6195db9839b1035a47697e19f2d2f17a4fcdc23bdc141eaa772db0d3"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.031140 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj" event={"ID":"812c885e-98f8-47cb-afd6-5fb537bea3e3","Type":"ContainerStarted","Data":"0fd5b382b8e06eafadcb03be8e9207051ecc27429ccd601d122940a68dbd42c4"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.047875 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q72ss" podStartSLOduration=164.047857396 podStartE2EDuration="2m44.047857396s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.046905028 +0000 UTC m=+226.944279638" watchObservedRunningTime="2026-03-18 00:11:49.047857396 +0000 UTC m=+226.945232006" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.062548 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" event={"ID":"f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a","Type":"ContainerStarted","Data":"8da75cf878c9d35faa1839be57796a0d25f7a9bd985707378a4b140456567fe9"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.066723 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" event={"ID":"da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8","Type":"ContainerStarted","Data":"38694c26e50acb7f02f924fdd57b2604e9532efdcab11bb42359641b64783bb2"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.069618 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.089381 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wxstd" event={"ID":"b876a42a-856b-4530-a601-3ca35db1078a","Type":"ContainerStarted","Data":"c3054825646b52dc765a806604c5e369e9c88b11ed3eadd87bc5c27ea8171966"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.105237 4983 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-9xkhq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" start-of-body= Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.105291 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" podUID="da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.106278 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" podStartSLOduration=165.106260489 podStartE2EDuration="2m45.106260489s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.103470196 +0000 UTC m=+227.000844806" watchObservedRunningTime="2026-03-18 00:11:49.106260489 +0000 UTC m=+227.003635099" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.145561 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:49 crc kubenswrapper[4983]: E0318 00:11:49.147333 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:49.647317524 +0000 UTC m=+227.544692124 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.152838 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" event={"ID":"ecc96f36-7ea7-4beb-977b-ae803b665b16","Type":"ContainerStarted","Data":"0b87ddbec5ebcfcabdcf713cde27b30478d75b4b49cbddf54a4df07702c75596"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.166052 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj" podStartSLOduration=165.166031833 podStartE2EDuration="2m45.166031833s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.13944871 +0000 UTC m=+227.036823320" watchObservedRunningTime="2026-03-18 00:11:49.166031833 +0000 UTC m=+227.063406443" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.183636 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-k424z" event={"ID":"29fc0965-6c9b-455f-b83e-f2fd1e04c6dc","Type":"ContainerStarted","Data":"639e5c6d3df49e6cc4c358695c5616e2baa87c0c0bd5078ba866861552a4d4f8"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.199218 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:11:49 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:11:49 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:11:49 crc kubenswrapper[4983]: healthz check failed Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.199265 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.212789 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wxstd" podStartSLOduration=164.212772518 podStartE2EDuration="2m44.212772518s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.177608048 +0000 UTC m=+227.074982668" watchObservedRunningTime="2026-03-18 00:11:49.212772518 +0000 UTC m=+227.110147128" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.219797 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-87bh5" event={"ID":"eb9a12c3-1119-4b1e-8d32-0a9696a00392","Type":"ContainerStarted","Data":"2857444edcbb15f915ac6bbfd8f5e625a4966f052892ed13d56af69146166e0b"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.223975 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-87bh5" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.240692 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" podStartSLOduration=164.24067569 podStartE2EDuration="2m44.24067569s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.213910152 +0000 UTC m=+227.111284772" watchObservedRunningTime="2026-03-18 00:11:49.24067569 +0000 UTC m=+227.138050300" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.242796 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" podStartSLOduration=164.242785073 podStartE2EDuration="2m44.242785073s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.240234167 +0000 UTC m=+227.137608797" watchObservedRunningTime="2026-03-18 00:11:49.242785073 +0000 UTC m=+227.140159683" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.256076 4983 patch_prober.go:28] interesting pod/downloads-7954f5f757-87bh5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.256130 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-87bh5" podUID="eb9a12c3-1119-4b1e-8d32-0a9696a00392" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.256660 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:49 crc kubenswrapper[4983]: E0318 00:11:49.257652 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:49.757629216 +0000 UTC m=+227.655003826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.281520 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" event={"ID":"eec36046-1daf-4f76-9a9e-3d9001830bb2","Type":"ContainerStarted","Data":"e3cc315202d934a5c47d8f0becd009d1be4a86f9d59ee0c327aed25ec8a2e939"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.286174 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" event={"ID":"84bde919-6dad-4080-977f-81cda3119a89","Type":"ContainerStarted","Data":"8a8382e27fa475a9c2b9bd983478414d2b152ef8df2c0fdc0858b0169d92cf20"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.287131 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" event={"ID":"7ddf406d-94c4-4ce9-b593-aebe0fca2b51","Type":"ContainerStarted","Data":"40177f0ab946debca3b68752da6222eeaa4247151b21d820c920bf4330aadfbc"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.288163 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.305264 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" event={"ID":"744c22bb-2f35-494b-b352-ff71691bd0af","Type":"ContainerStarted","Data":"5468285ed9a036b1f627399c456f170e5c3a6819bb88d047a1035d2130e0eab1"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.305583 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.309522 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" event={"ID":"71e981d5-d657-46ef-9ec0-4ddad58be6ee","Type":"ContainerStarted","Data":"ab02e9513fbca003c2cfae1f26cd924a07971a7c553f02c70c481bfcd9fdc01a"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.312602 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hmcxg" podStartSLOduration=165.312581666 podStartE2EDuration="2m45.312581666s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.311654479 +0000 UTC m=+227.209029089" watchObservedRunningTime="2026-03-18 00:11:49.312581666 +0000 UTC m=+227.209956276" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.312779 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-87bh5" podStartSLOduration=165.312775072 podStartE2EDuration="2m45.312775072s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.279437397 +0000 UTC m=+227.176812017" watchObservedRunningTime="2026-03-18 00:11:49.312775072 +0000 UTC m=+227.210149682" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.317690 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.325895 4983 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-8qf4k container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.325955 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" podUID="7ddf406d-94c4-4ce9-b593-aebe0fca2b51" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.349706 4983 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rpftn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.349764 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" podUID="71e981d5-d657-46ef-9ec0-4ddad58be6ee" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.350922 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" podStartSLOduration=164.35091181 podStartE2EDuration="2m44.35091181s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.348521589 +0000 UTC m=+227.245896219" watchObservedRunningTime="2026-03-18 00:11:49.35091181 +0000 UTC m=+227.248286420" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.360567 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:49 crc kubenswrapper[4983]: E0318 00:11:49.361413 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:49.861400173 +0000 UTC m=+227.758774783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.362812 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" event={"ID":"d3f900da-08ff-4983-89b7-4dea2613f0fd","Type":"ContainerStarted","Data":"5a9c188a1e644d2042f511fbc05fa85b68c67ad8871ce75b8a847447dfe5343c"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.395239 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" event={"ID":"1321ddb2-5769-482f-b4dc-1971dad30fdf","Type":"ContainerStarted","Data":"689b7e591fd6dced9ce7ecc04d44009ec080da8e1855a84313f874b36cc3e804"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.396160 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.398772 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" podStartSLOduration=165.398762628 podStartE2EDuration="2m45.398762628s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.397028326 +0000 UTC m=+227.294402936" watchObservedRunningTime="2026-03-18 00:11:49.398762628 +0000 UTC m=+227.296137238" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.414916 4983 generic.go:334] "Generic (PLEG): container finished" podID="0774a9cb-d9cb-45a2-89e1-384902cf8f94" containerID="96e054ab20b2a20022435e1fa3b784ceac9a696a7bb579a935a4f9a02a7a2f90" exitCode=0 Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.415017 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" event={"ID":"0774a9cb-d9cb-45a2-89e1-384902cf8f94","Type":"ContainerDied","Data":"96e054ab20b2a20022435e1fa3b784ceac9a696a7bb579a935a4f9a02a7a2f90"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.435040 4983 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-wsk9g container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.435323 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" podUID="1321ddb2-5769-482f-b4dc-1971dad30fdf" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.452419 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" event={"ID":"eaebd031-fb05-489b-8be1-e4e115bdcb7d","Type":"ContainerStarted","Data":"b31da3b26132d306119575f78e1ff308cd1cc90f8e0207ec341800deb3b6cdc4"} Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.455687 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" podStartSLOduration=164.455671547 podStartE2EDuration="2m44.455671547s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.442076371 +0000 UTC m=+227.339450991" watchObservedRunningTime="2026-03-18 00:11:49.455671547 +0000 UTC m=+227.353046157" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.469698 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.470146 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:49 crc kubenswrapper[4983]: E0318 00:11:49.470964 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:49.970952152 +0000 UTC m=+227.868326762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.484257 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.487257 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" podStartSLOduration=164.487243919 podStartE2EDuration="2m44.487243919s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.486369503 +0000 UTC m=+227.383744123" watchObservedRunningTime="2026-03-18 00:11:49.487243919 +0000 UTC m=+227.384618529" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.514731 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-gx8mh" podStartSLOduration=164.514713578 podStartE2EDuration="2m44.514713578s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.513273765 +0000 UTC m=+227.410648375" watchObservedRunningTime="2026-03-18 00:11:49.514713578 +0000 UTC m=+227.412088188" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.571523 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:49 crc kubenswrapper[4983]: E0318 00:11:49.579211 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.079195613 +0000 UTC m=+227.976570223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.645038 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" podStartSLOduration=164.645022777 podStartE2EDuration="2m44.645022777s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.644615965 +0000 UTC m=+227.541990575" watchObservedRunningTime="2026-03-18 00:11:49.645022777 +0000 UTC m=+227.542397387" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.673706 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" podStartSLOduration=164.673690123 podStartE2EDuration="2m44.673690123s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:49.67158199 +0000 UTC m=+227.568956610" watchObservedRunningTime="2026-03-18 00:11:49.673690123 +0000 UTC m=+227.571064733" Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.674028 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:49 crc kubenswrapper[4983]: E0318 00:11:49.674401 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.174384773 +0000 UTC m=+228.071759383 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.775267 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:49 crc kubenswrapper[4983]: E0318 00:11:49.775749 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.275736358 +0000 UTC m=+228.173110968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.876470 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:49 crc kubenswrapper[4983]: E0318 00:11:49.876676 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.376647019 +0000 UTC m=+228.274021619 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.876719 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:49 crc kubenswrapper[4983]: E0318 00:11:49.877058 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.377046681 +0000 UTC m=+228.274421291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:49 crc kubenswrapper[4983]: I0318 00:11:49.977800 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:49 crc kubenswrapper[4983]: E0318 00:11:49.978299 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.478279702 +0000 UTC m=+228.375654312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.079596 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:50 crc kubenswrapper[4983]: E0318 00:11:50.080070 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.580054709 +0000 UTC m=+228.477429319 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.180352 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:50 crc kubenswrapper[4983]: E0318 00:11:50.180481 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.680461936 +0000 UTC m=+228.577836546 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.180654 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:50 crc kubenswrapper[4983]: E0318 00:11:50.180968 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.680956511 +0000 UTC m=+228.578331121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.198388 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:11:50 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:11:50 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:11:50 crc kubenswrapper[4983]: healthz check failed Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.198440 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.282209 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:50 crc kubenswrapper[4983]: E0318 00:11:50.282513 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.78248552 +0000 UTC m=+228.679860130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.282709 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:50 crc kubenswrapper[4983]: E0318 00:11:50.283062 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.783049957 +0000 UTC m=+228.680424557 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.383465 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:50 crc kubenswrapper[4983]: E0318 00:11:50.383627 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.883605178 +0000 UTC m=+228.780979788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.383812 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:50 crc kubenswrapper[4983]: E0318 00:11:50.384148 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.884140354 +0000 UTC m=+228.781514964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.464067 4983 patch_prober.go:28] interesting pod/console-operator-58897d9998-slqc5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": context deadline exceeded" start-of-body= Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.464132 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-slqc5" podUID="04baa777-3a33-4b5a-9808-99a84b1b8005" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": context deadline exceeded" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.481193 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" event={"ID":"f3b59e54-a961-4c4b-8b4a-ddbd24c5aa5a","Type":"ContainerStarted","Data":"e3a199888471c049c708edd08ca626a8085a8f77e7132b53b704c46ad70c7ba9"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.484968 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:50 crc kubenswrapper[4983]: E0318 00:11:50.485306 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:50.985289873 +0000 UTC m=+228.882664483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.488998 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" event={"ID":"1321ddb2-5769-482f-b4dc-1971dad30fdf","Type":"ContainerStarted","Data":"c9ddcf6ee4708ad7ed71485d6d7b18edfe084d96a4a04db83838e9be59104984"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.489448 4983 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-wsk9g container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.489486 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" podUID="1321ddb2-5769-482f-b4dc-1971dad30fdf" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.494569 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" event={"ID":"0774a9cb-d9cb-45a2-89e1-384902cf8f94","Type":"ContainerStarted","Data":"d23bdb488301b25b17a332b9025d82aa7bea89ff91076b4ef7f5b227077435ed"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.496245 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" event={"ID":"49880d65-bff4-45d1-8fb7-907e524568fc","Type":"ContainerStarted","Data":"0a531f558ba6bbe316bbb57d54a13990d28c32f8b347bf52bbc6046a45f020fa"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.497306 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" event={"ID":"602912b1-d12a-406a-ba6e-e0d94dc62038","Type":"ContainerStarted","Data":"21093baebef5e62832dacba00ecd7e93e6fd4fd43659782d70b3341bf1eb3209"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.511204 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kf9gt" event={"ID":"bd950f1d-4d97-4ed4-8d9c-04245019bbdb","Type":"ContainerStarted","Data":"262355760da727bf370302adda4d2bd326e8354173e4555564da480947d6e472"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.514179 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h74lp" event={"ID":"53bc5818-4983-4f5f-8a3c-edc02aff5325","Type":"ContainerStarted","Data":"2109d18f61bf0d43e472c0ec650d6d83ca928a5b77d821aecc688beb8aea654b"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.514206 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h74lp" event={"ID":"53bc5818-4983-4f5f-8a3c-edc02aff5325","Type":"ContainerStarted","Data":"90efcb35198956da7eac001978b566358dafc5c2cdb72cd373da5159ea5612ee"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.519439 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" event={"ID":"d7237321-db39-4ed6-a22e-b3b57c5b6be6","Type":"ContainerStarted","Data":"39f7c1bd70b532ef0f4fdffd40721197badfa2757630812044f97f0a5a20f06b"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.519483 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" event={"ID":"d7237321-db39-4ed6-a22e-b3b57c5b6be6","Type":"ContainerStarted","Data":"f781db06cd54fface53a13e7441f178e4a725e16e06a25675a73953af0a6def2"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.519559 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.522745 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w8wsj" event={"ID":"812c885e-98f8-47cb-afd6-5fb537bea3e3","Type":"ContainerStarted","Data":"625d6f18bb4ada3fbe56c6fc16dce8124649fd60f609cb02626efded260878d7"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.525645 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-f8svz" podStartSLOduration=165.525634157 podStartE2EDuration="2m45.525634157s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:50.52406418 +0000 UTC m=+228.421438800" watchObservedRunningTime="2026-03-18 00:11:50.525634157 +0000 UTC m=+228.423008767" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.526064 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9h7fc" event={"ID":"eaebd031-fb05-489b-8be1-e4e115bdcb7d","Type":"ContainerStarted","Data":"cb5c646e91c2a09f587fa7847711981e0c85175665942d160f5d130655c8a661"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.538666 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-k424z" event={"ID":"29fc0965-6c9b-455f-b83e-f2fd1e04c6dc","Type":"ContainerStarted","Data":"f900bc0a6e618dfa0db4dab5284118b806986af3a1e90a3f3083a8620a28f9b2"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.538798 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-k424z" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.550131 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b8gtz" event={"ID":"63bc8988-a69a-4385-8ff3-45edafde495a","Type":"ContainerStarted","Data":"3e38c4622a4a1757add825d5bfbb54cc06ba29cd49893a36dabb34034d983ec7"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.552388 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" event={"ID":"a57b48db-a186-4a73-b139-a9249a19a3ed","Type":"ContainerStarted","Data":"c19c4183a62f58ccb3f1684a63a7232315646fe7b86f7d202b5509be50b64552"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.555509 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-99dq7" podStartSLOduration=165.555493858 podStartE2EDuration="2m45.555493858s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:50.555458077 +0000 UTC m=+228.452832697" watchObservedRunningTime="2026-03-18 00:11:50.555493858 +0000 UTC m=+228.452868468" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.565494 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-85cfx" event={"ID":"6d53acd4-182c-4dd4-93e1-bccac2a0a522","Type":"ContainerStarted","Data":"f3284de123dcc02b14ed6d77fb95e181b5dbee02f2ed8e05e7ee3bbd15d04b90"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.565535 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-85cfx" event={"ID":"6d53acd4-182c-4dd4-93e1-bccac2a0a522","Type":"ContainerStarted","Data":"36659d31528244e50ac64e17b6f85784fbded6ed019f1f08699b460a499ff103"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.571821 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" event={"ID":"84bde919-6dad-4080-977f-81cda3119a89","Type":"ContainerStarted","Data":"a3985da7dc212d978ed2630aef0e9afa5cd9d974e7d3003c75653771336c3ed3"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.581190 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" event={"ID":"7ddf406d-94c4-4ce9-b593-aebe0fca2b51","Type":"ContainerStarted","Data":"dbb5410ed54936cb3c123f91e8626d719704b9ed1ebed51431771d348f4a7feb"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.584154 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kqxb6" event={"ID":"ecc96f36-7ea7-4beb-977b-ae803b665b16","Type":"ContainerStarted","Data":"ccb0ecda51d5fd3bd3311a898b5157cf8eaab6029412702945cc2bff4958e906"} Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.586272 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.587296 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.587449 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.589366 4983 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-9xkhq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" start-of-body= Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.589419 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" podUID="da1adbd9-c3f4-4db6-84cd-fe7b9629c4a8" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.589750 4983 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rpftn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.589766 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" podUID="71e981d5-d657-46ef-9ec0-4ddad58be6ee" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.589850 4983 patch_prober.go:28] interesting pod/downloads-7954f5f757-87bh5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.589906 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-87bh5" podUID="eb9a12c3-1119-4b1e-8d32-0a9696a00392" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Mar 18 00:11:50 crc kubenswrapper[4983]: E0318 00:11:50.590964 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:51.090946476 +0000 UTC m=+228.988321086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.593130 4983 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-8qf4k container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.593495 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" podUID="7ddf406d-94c4-4ce9-b593-aebe0fca2b51" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.623086 4983 patch_prober.go:28] interesting pod/apiserver-76f77b778f-h74lp container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.623154 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-h74lp" podUID="53bc5818-4983-4f5f-8a3c-edc02aff5325" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.654788 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-h74lp" podStartSLOduration=166.65477426 podStartE2EDuration="2m46.65477426s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:50.654216514 +0000 UTC m=+228.551591144" watchObservedRunningTime="2026-03-18 00:11:50.65477426 +0000 UTC m=+228.552148870" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.656070 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" podStartSLOduration=165.656064309 podStartE2EDuration="2m45.656064309s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:50.608645454 +0000 UTC m=+228.506020064" watchObservedRunningTime="2026-03-18 00:11:50.656064309 +0000 UTC m=+228.553438919" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.687623 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:50 crc kubenswrapper[4983]: E0318 00:11:50.691705 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:51.191680622 +0000 UTC m=+229.089055232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.701509 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-b8gtz" podStartSLOduration=7.701488584 podStartE2EDuration="7.701488584s" podCreationTimestamp="2026-03-18 00:11:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:50.698140565 +0000 UTC m=+228.595515175" watchObservedRunningTime="2026-03-18 00:11:50.701488584 +0000 UTC m=+228.598863184" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.729477 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-k424z" podStartSLOduration=7.729458349 podStartE2EDuration="7.729458349s" podCreationTimestamp="2026-03-18 00:11:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:50.727081628 +0000 UTC m=+228.624456258" watchObservedRunningTime="2026-03-18 00:11:50.729458349 +0000 UTC m=+228.626832959" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.750147 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zsklf" podStartSLOduration=165.750130396 podStartE2EDuration="2m45.750130396s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:50.748897369 +0000 UTC m=+228.646271979" watchObservedRunningTime="2026-03-18 00:11:50.750130396 +0000 UTC m=+228.647505006" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.790037 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.790206 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-w4hpj" podStartSLOduration=165.790164431 podStartE2EDuration="2m45.790164431s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:50.789095179 +0000 UTC m=+228.686469799" watchObservedRunningTime="2026-03-18 00:11:50.790164431 +0000 UTC m=+228.687539051" Mar 18 00:11:50 crc kubenswrapper[4983]: E0318 00:11:50.790479 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:51.290456989 +0000 UTC m=+229.187831599 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.817141 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-85cfx" podStartSLOduration=165.817125625 podStartE2EDuration="2m45.817125625s" podCreationTimestamp="2026-03-18 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:50.815702773 +0000 UTC m=+228.713077383" watchObservedRunningTime="2026-03-18 00:11:50.817125625 +0000 UTC m=+228.714500225" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.844028 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.844559 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.853808 4983 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-g7fq4 container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.853878 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" podUID="0774a9cb-d9cb-45a2-89e1-384902cf8f94" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.893061 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:50 crc kubenswrapper[4983]: E0318 00:11:50.893425 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:51.393405362 +0000 UTC m=+229.290779982 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.952870 4983 ???:1] "http: TLS handshake error from 192.168.126.11:59850: no serving certificate available for the kubelet" Mar 18 00:11:50 crc kubenswrapper[4983]: I0318 00:11:50.995045 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:50 crc kubenswrapper[4983]: E0318 00:11:50.995844 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:51.495819558 +0000 UTC m=+229.393194158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.028069 4983 ???:1] "http: TLS handshake error from 192.168.126.11:59862: no serving certificate available for the kubelet" Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.096345 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.096585 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:51.596540954 +0000 UTC m=+229.493915564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.096654 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.097120 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:51.59710279 +0000 UTC m=+229.494477400 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.122913 4983 ???:1] "http: TLS handshake error from 192.168.126.11:59868: no serving certificate available for the kubelet" Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.195790 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:11:51 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:11:51 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:11:51 crc kubenswrapper[4983]: healthz check failed Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.195881 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.197305 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.197693 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:51.697675322 +0000 UTC m=+229.595049932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.229061 4983 ???:1] "http: TLS handshake error from 192.168.126.11:59882: no serving certificate available for the kubelet" Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.298382 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.298729 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:51.798718067 +0000 UTC m=+229.696092667 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.334262 4983 ???:1] "http: TLS handshake error from 192.168.126.11:59892: no serving certificate available for the kubelet" Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.379265 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zlqph"] Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.405212 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.405399 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:51.90537259 +0000 UTC m=+229.802747200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.405484 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.405761 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:51.905745771 +0000 UTC m=+229.803120381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.419978 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4"] Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.444606 4983 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-pg4wt container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.444662 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" podUID="744c22bb-2f35-494b-b352-ff71691bd0af" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.444716 4983 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-pg4wt container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.444775 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" podUID="744c22bb-2f35-494b-b352-ff71691bd0af" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.474457 4983 ???:1] "http: TLS handshake error from 192.168.126.11:59896: no serving certificate available for the kubelet" Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.506833 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.507045 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.007002613 +0000 UTC m=+229.904377223 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.507081 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.507407 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.007390854 +0000 UTC m=+229.904765464 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.607671 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.607899 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.107877293 +0000 UTC m=+230.005251913 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.607995 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.608299 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.108290155 +0000 UTC m=+230.005664775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.613178 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kzks7" event={"ID":"799072d7-3ce6-4742-be4c-065ecea7e24c","Type":"ContainerStarted","Data":"1838e5bd2c4b06d85e921de6e7a99f3010da638e8667b7f95063a6fb4f1ea055"} Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.613315 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" podUID="5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d" containerName="route-controller-manager" containerID="cri-o://d9eb68e8bcc8979fda5ff8f269ddbb7bc410310bba6eba0fe60bd4b384c28556" gracePeriod=30 Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.614831 4983 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-8qf4k container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.614858 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" podUID="7ddf406d-94c4-4ce9-b593-aebe0fca2b51" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.614893 4983 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-wsk9g container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.614968 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" podUID="1321ddb2-5769-482f-b4dc-1971dad30fdf" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.615208 4983 patch_prober.go:28] interesting pod/downloads-7954f5f757-87bh5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.615253 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-87bh5" podUID="eb9a12c3-1119-4b1e-8d32-0a9696a00392" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.615321 4983 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rpftn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.615342 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" podUID="71e981d5-d657-46ef-9ec0-4ddad58be6ee" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.616537 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" podUID="0b2013fc-d64f-4f21-be62-e17f3855f6c1" containerName="controller-manager" containerID="cri-o://3cb762f5e037bbf892d7687651e70656cb61ca252d75cac1cc7887be5bad691b" gracePeriod=30 Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.671508 4983 ???:1] "http: TLS handshake error from 192.168.126.11:59904: no serving certificate available for the kubelet" Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.709076 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.709232 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.209204457 +0000 UTC m=+230.106579067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.709752 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.711298 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.211281699 +0000 UTC m=+230.108656309 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.811258 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.811441 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.311412847 +0000 UTC m=+230.208787457 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.811889 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.812213 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.312201161 +0000 UTC m=+230.209575771 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.913106 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.913401 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.413344159 +0000 UTC m=+230.310718789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:51 crc kubenswrapper[4983]: I0318 00:11:51.913462 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:51 crc kubenswrapper[4983]: E0318 00:11:51.913771 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.413759251 +0000 UTC m=+230.311133851 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.014629 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.014731 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.514714334 +0000 UTC m=+230.412088944 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.015033 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.015342 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.515333672 +0000 UTC m=+230.412708282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.015816 4983 ???:1] "http: TLS handshake error from 192.168.126.11:59912: no serving certificate available for the kubelet" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.116482 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.116684 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.616658686 +0000 UTC m=+230.514033296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.116752 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.117099 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.617091159 +0000 UTC m=+230.514465769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.194412 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:11:52 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:11:52 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:11:52 crc kubenswrapper[4983]: healthz check failed Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.194479 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.217689 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.218027 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.718010581 +0000 UTC m=+230.615385191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.299286 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xkhq" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.318841 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.319187 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.819175399 +0000 UTC m=+230.716550009 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.433291 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.433644 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:52.933627454 +0000 UTC m=+230.831002054 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.537641 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.538007 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:53.037994909 +0000 UTC m=+230.935369519 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.623478 4983 generic.go:334] "Generic (PLEG): container finished" podID="5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d" containerID="d9eb68e8bcc8979fda5ff8f269ddbb7bc410310bba6eba0fe60bd4b384c28556" exitCode=0 Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.623713 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" event={"ID":"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d","Type":"ContainerDied","Data":"d9eb68e8bcc8979fda5ff8f269ddbb7bc410310bba6eba0fe60bd4b384c28556"} Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.623742 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" event={"ID":"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d","Type":"ContainerDied","Data":"94a5ec1ff11e094545a5a0e4cb15ca88ee65b0d0ec1cb036c60c193b827e180a"} Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.623753 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94a5ec1ff11e094545a5a0e4cb15ca88ee65b0d0ec1cb036c60c193b827e180a" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.625398 4983 generic.go:334] "Generic (PLEG): container finished" podID="0b2013fc-d64f-4f21-be62-e17f3855f6c1" containerID="3cb762f5e037bbf892d7687651e70656cb61ca252d75cac1cc7887be5bad691b" exitCode=0 Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.625509 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" event={"ID":"0b2013fc-d64f-4f21-be62-e17f3855f6c1","Type":"ContainerDied","Data":"3cb762f5e037bbf892d7687651e70656cb61ca252d75cac1cc7887be5bad691b"} Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.634219 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.638489 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.638900 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:53.13887895 +0000 UTC m=+231.036253560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.685996 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4"] Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.686254 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d" containerName="route-controller-manager" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.686272 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d" containerName="route-controller-manager" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.686391 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d" containerName="route-controller-manager" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.686709 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4"] Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.686787 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.694533 4983 ???:1] "http: TLS handshake error from 192.168.126.11:59918: no serving certificate available for the kubelet" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.761301 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-client-ca\") pod \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.761360 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-serving-cert\") pod \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.761406 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-config\") pod \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.761558 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf67f\" (UniqueName: \"kubernetes.io/projected/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-kube-api-access-nf67f\") pod \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\" (UID: \"5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d\") " Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.761688 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn8ff\" (UniqueName: \"kubernetes.io/projected/12aae035-1b1b-4e29-8866-1b3fde863816-kube-api-access-nn8ff\") pod \"route-controller-manager-84b4fdf79c-z4cv4\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.761738 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12aae035-1b1b-4e29-8866-1b3fde863816-config\") pod \"route-controller-manager-84b4fdf79c-z4cv4\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.761793 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.761969 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12aae035-1b1b-4e29-8866-1b3fde863816-serving-cert\") pod \"route-controller-manager-84b4fdf79c-z4cv4\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.762007 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12aae035-1b1b-4e29-8866-1b3fde863816-client-ca\") pod \"route-controller-manager-84b4fdf79c-z4cv4\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.762531 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-config" (OuterVolumeSpecName: "config") pod "5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d" (UID: "5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.762823 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:53.262806718 +0000 UTC m=+231.160181408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.763262 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-client-ca" (OuterVolumeSpecName: "client-ca") pod "5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d" (UID: "5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.779198 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d" (UID: "5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.781498 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-kube-api-access-nf67f" (OuterVolumeSpecName: "kube-api-access-nf67f") pod "5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d" (UID: "5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d"). InnerVolumeSpecName "kube-api-access-nf67f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.822775 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.866475 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.866713 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12aae035-1b1b-4e29-8866-1b3fde863816-serving-cert\") pod \"route-controller-manager-84b4fdf79c-z4cv4\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.866736 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12aae035-1b1b-4e29-8866-1b3fde863816-client-ca\") pod \"route-controller-manager-84b4fdf79c-z4cv4\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.866765 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn8ff\" (UniqueName: \"kubernetes.io/projected/12aae035-1b1b-4e29-8866-1b3fde863816-kube-api-access-nn8ff\") pod \"route-controller-manager-84b4fdf79c-z4cv4\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.866790 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12aae035-1b1b-4e29-8866-1b3fde863816-config\") pod \"route-controller-manager-84b4fdf79c-z4cv4\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.866839 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.866849 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.866857 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.866866 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf67f\" (UniqueName: \"kubernetes.io/projected/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d-kube-api-access-nf67f\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.867888 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12aae035-1b1b-4e29-8866-1b3fde863816-config\") pod \"route-controller-manager-84b4fdf79c-z4cv4\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.867984 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:53.367967506 +0000 UTC m=+231.265342116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.868662 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12aae035-1b1b-4e29-8866-1b3fde863816-client-ca\") pod \"route-controller-manager-84b4fdf79c-z4cv4\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.883043 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12aae035-1b1b-4e29-8866-1b3fde863816-serving-cert\") pod \"route-controller-manager-84b4fdf79c-z4cv4\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.903334 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn8ff\" (UniqueName: \"kubernetes.io/projected/12aae035-1b1b-4e29-8866-1b3fde863816-kube-api-access-nn8ff\") pod \"route-controller-manager-84b4fdf79c-z4cv4\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.932005 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4"] Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.932428 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.948369 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-57d94b7d77-crjwh"] Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.948566 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2013fc-d64f-4f21-be62-e17f3855f6c1" containerName="controller-manager" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.948578 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2013fc-d64f-4f21-be62-e17f3855f6c1" containerName="controller-manager" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.948680 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2013fc-d64f-4f21-be62-e17f3855f6c1" containerName="controller-manager" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.949118 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.968427 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq7j2\" (UniqueName: \"kubernetes.io/projected/0b2013fc-d64f-4f21-be62-e17f3855f6c1-kube-api-access-lq7j2\") pod \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.968473 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-proxy-ca-bundles\") pod \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.968503 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b2013fc-d64f-4f21-be62-e17f3855f6c1-serving-cert\") pod \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.968550 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-config\") pod \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.968605 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-client-ca\") pod \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\" (UID: \"0b2013fc-d64f-4f21-be62-e17f3855f6c1\") " Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.968851 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:52 crc kubenswrapper[4983]: E0318 00:11:52.969231 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:53.469216178 +0000 UTC m=+231.366590788 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.970440 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-config" (OuterVolumeSpecName: "config") pod "0b2013fc-d64f-4f21-be62-e17f3855f6c1" (UID: "0b2013fc-d64f-4f21-be62-e17f3855f6c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.970516 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0b2013fc-d64f-4f21-be62-e17f3855f6c1" (UID: "0b2013fc-d64f-4f21-be62-e17f3855f6c1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.971415 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-client-ca" (OuterVolumeSpecName: "client-ca") pod "0b2013fc-d64f-4f21-be62-e17f3855f6c1" (UID: "0b2013fc-d64f-4f21-be62-e17f3855f6c1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.975343 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2013fc-d64f-4f21-be62-e17f3855f6c1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b2013fc-d64f-4f21-be62-e17f3855f6c1" (UID: "0b2013fc-d64f-4f21-be62-e17f3855f6c1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.976348 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57d94b7d77-crjwh"] Mar 18 00:11:52 crc kubenswrapper[4983]: I0318 00:11:52.980091 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b2013fc-d64f-4f21-be62-e17f3855f6c1-kube-api-access-lq7j2" (OuterVolumeSpecName: "kube-api-access-lq7j2") pod "0b2013fc-d64f-4f21-be62-e17f3855f6c1" (UID: "0b2013fc-d64f-4f21-be62-e17f3855f6c1"). InnerVolumeSpecName "kube-api-access-lq7j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.069450 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.069697 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-proxy-ca-bundles\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.069738 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c27c1f03-b245-4beb-9de4-f17cb0844f7a-serving-cert\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.069772 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9q8l\" (UniqueName: \"kubernetes.io/projected/c27c1f03-b245-4beb-9de4-f17cb0844f7a-kube-api-access-v9q8l\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.069803 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-config\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.069864 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-client-ca\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.069973 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.069993 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.070003 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq7j2\" (UniqueName: \"kubernetes.io/projected/0b2013fc-d64f-4f21-be62-e17f3855f6c1-kube-api-access-lq7j2\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.070020 4983 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0b2013fc-d64f-4f21-be62-e17f3855f6c1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.070029 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b2013fc-d64f-4f21-be62-e17f3855f6c1-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:53 crc kubenswrapper[4983]: E0318 00:11:53.070140 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:53.570121929 +0000 UTC m=+231.467496539 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.171240 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c27c1f03-b245-4beb-9de4-f17cb0844f7a-serving-cert\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.171516 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9q8l\" (UniqueName: \"kubernetes.io/projected/c27c1f03-b245-4beb-9de4-f17cb0844f7a-kube-api-access-v9q8l\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.171645 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-config\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.171768 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.171866 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-client-ca\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.172007 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-proxy-ca-bundles\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: E0318 00:11:53.172978 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:53.672958588 +0000 UTC m=+231.570333198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.173364 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-config\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.174148 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-client-ca\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.175046 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-proxy-ca-bundles\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.178275 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c27c1f03-b245-4beb-9de4-f17cb0844f7a-serving-cert\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.190347 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9q8l\" (UniqueName: \"kubernetes.io/projected/c27c1f03-b245-4beb-9de4-f17cb0844f7a-kube-api-access-v9q8l\") pod \"controller-manager-57d94b7d77-crjwh\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.235471 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:11:53 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:11:53 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:11:53 crc kubenswrapper[4983]: healthz check failed Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.235855 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.274659 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.275229 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:53 crc kubenswrapper[4983]: E0318 00:11:53.275725 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:53.775704314 +0000 UTC m=+231.673078924 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.376783 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:53 crc kubenswrapper[4983]: E0318 00:11:53.377124 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:53.87711057 +0000 UTC m=+231.774485180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.415562 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4"] Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.478171 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:53 crc kubenswrapper[4983]: E0318 00:11:53.478377 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:53.978349041 +0000 UTC m=+231.875723651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.478479 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:53 crc kubenswrapper[4983]: E0318 00:11:53.478835 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:53.978825405 +0000 UTC m=+231.876200015 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.566632 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57d94b7d77-crjwh"] Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.579662 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:53 crc kubenswrapper[4983]: E0318 00:11:53.580021 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:54.080003225 +0000 UTC m=+231.977377835 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.646316 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" event={"ID":"c27c1f03-b245-4beb-9de4-f17cb0844f7a","Type":"ContainerStarted","Data":"6dbbad3fd96b6fb449021e881645bee81a8cbca2256b558a8b2fdd2f9ad0778a"} Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.658320 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" event={"ID":"0b2013fc-d64f-4f21-be62-e17f3855f6c1","Type":"ContainerDied","Data":"ef176c7de73ae407113e93e01247788d708a5b67f4bfc39d062abeea19177d96"} Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.658374 4983 scope.go:117] "RemoveContainer" containerID="3cb762f5e037bbf892d7687651e70656cb61ca252d75cac1cc7887be5bad691b" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.658537 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zlqph" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.664442 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.664992 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" event={"ID":"12aae035-1b1b-4e29-8866-1b3fde863816","Type":"ContainerStarted","Data":"25eb8c6043db7ff2bff998648ec50b69c15942f4e0ad1fb2bdd0cc82e67399fc"} Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.684682 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.687617 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4"] Mar 18 00:11:53 crc kubenswrapper[4983]: E0318 00:11:53.688204 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:54.188157602 +0000 UTC m=+232.085532212 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.690694 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5j7r4"] Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.706231 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zlqph"] Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.710147 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zlqph"] Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.786505 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:53 crc kubenswrapper[4983]: E0318 00:11:53.786886 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:54.286865169 +0000 UTC m=+232.184239789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.890526 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:53 crc kubenswrapper[4983]: E0318 00:11:53.890885 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:54.390868882 +0000 UTC m=+232.288243492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.991674 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:53 crc kubenswrapper[4983]: E0318 00:11:53.991876 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:54.491848146 +0000 UTC m=+232.389222756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:53 crc kubenswrapper[4983]: I0318 00:11:53.992328 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:53 crc kubenswrapper[4983]: E0318 00:11:53.992716 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:54.492700501 +0000 UTC m=+232.390075111 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.004541 4983 ???:1] "http: TLS handshake error from 192.168.126.11:59926: no serving certificate available for the kubelet" Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.093461 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:54 crc kubenswrapper[4983]: E0318 00:11:54.093863 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:54.593839999 +0000 UTC m=+232.491214609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.194671 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:54 crc kubenswrapper[4983]: E0318 00:11:54.195127 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:54.695111871 +0000 UTC m=+232.592486481 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.195380 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:11:54 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:11:54 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:11:54 crc kubenswrapper[4983]: healthz check failed Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.195415 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.295866 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:54 crc kubenswrapper[4983]: E0318 00:11:54.296205 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:54.796187148 +0000 UTC m=+232.693561758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.397435 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:54 crc kubenswrapper[4983]: E0318 00:11:54.397741 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:54.897729378 +0000 UTC m=+232.795103988 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.451399 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pg4wt" Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.498543 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:54 crc kubenswrapper[4983]: E0318 00:11:54.498951 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:54.998915198 +0000 UTC m=+232.896289808 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.600014 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:54 crc kubenswrapper[4983]: E0318 00:11:54.601290 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:55.101276123 +0000 UTC m=+232.998650723 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.674199 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" event={"ID":"c27c1f03-b245-4beb-9de4-f17cb0844f7a","Type":"ContainerStarted","Data":"d72721315f30ae7d13f4a96c5cc25b8b3599a375f273fcdcedce84e09b1c0249"} Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.675265 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.683720 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.684221 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" event={"ID":"12aae035-1b1b-4e29-8866-1b3fde863816","Type":"ContainerStarted","Data":"e285577cdd159cab74668fd228651187aec0ba78385e0118987b2db96c0ff114"} Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.684334 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" podUID="12aae035-1b1b-4e29-8866-1b3fde863816" containerName="route-controller-manager" containerID="cri-o://e285577cdd159cab74668fd228651187aec0ba78385e0118987b2db96c0ff114" gracePeriod=30 Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.684632 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.692496 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" podStartSLOduration=2.692479564 podStartE2EDuration="2.692479564s" podCreationTimestamp="2026-03-18 00:11:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:54.692299759 +0000 UTC m=+232.589674369" watchObservedRunningTime="2026-03-18 00:11:54.692479564 +0000 UTC m=+232.589854174" Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.693313 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.700701 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:54 crc kubenswrapper[4983]: E0318 00:11:54.701036 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:55.201020849 +0000 UTC m=+233.098395459 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.710764 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" podStartSLOduration=3.7107427189999997 podStartE2EDuration="3.710742719s" podCreationTimestamp="2026-03-18 00:11:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:11:54.709306557 +0000 UTC m=+232.606681167" watchObservedRunningTime="2026-03-18 00:11:54.710742719 +0000 UTC m=+232.608117329" Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.802689 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:54 crc kubenswrapper[4983]: E0318 00:11:54.803772 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:55.303759755 +0000 UTC m=+233.201134365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.851856 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b2013fc-d64f-4f21-be62-e17f3855f6c1" path="/var/lib/kubelet/pods/0b2013fc-d64f-4f21-be62-e17f3855f6c1/volumes" Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.852500 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d" path="/var/lib/kubelet/pods/5c4f3aae-cfe4-4935-b1d6-3c7be1557e5d/volumes" Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.904751 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:54 crc kubenswrapper[4983]: E0318 00:11:54.905037 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:55.404988776 +0000 UTC m=+233.302363386 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:54 crc kubenswrapper[4983]: I0318 00:11:54.905530 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:54 crc kubenswrapper[4983]: E0318 00:11:54.905963 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:55.405945725 +0000 UTC m=+233.303320335 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.006651 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.006859 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:55.506828565 +0000 UTC m=+233.404203175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.006962 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.007294 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:55.507281589 +0000 UTC m=+233.404656199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.107436 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.107838 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.108225 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:55.6081815 +0000 UTC m=+233.505556110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.109659 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.110074 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:55.610064876 +0000 UTC m=+233.507439486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.140392 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb"] Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.140646 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12aae035-1b1b-4e29-8866-1b3fde863816" containerName="route-controller-manager" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.140663 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="12aae035-1b1b-4e29-8866-1b3fde863816" containerName="route-controller-manager" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.140781 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="12aae035-1b1b-4e29-8866-1b3fde863816" containerName="route-controller-manager" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.141246 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.143256 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb"] Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.193562 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:11:55 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:11:55 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:11:55 crc kubenswrapper[4983]: healthz check failed Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.193631 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.210723 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12aae035-1b1b-4e29-8866-1b3fde863816-config\") pod \"12aae035-1b1b-4e29-8866-1b3fde863816\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.210813 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12aae035-1b1b-4e29-8866-1b3fde863816-client-ca\") pod \"12aae035-1b1b-4e29-8866-1b3fde863816\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.210840 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn8ff\" (UniqueName: \"kubernetes.io/projected/12aae035-1b1b-4e29-8866-1b3fde863816-kube-api-access-nn8ff\") pod \"12aae035-1b1b-4e29-8866-1b3fde863816\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.210881 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12aae035-1b1b-4e29-8866-1b3fde863816-serving-cert\") pod \"12aae035-1b1b-4e29-8866-1b3fde863816\" (UID: \"12aae035-1b1b-4e29-8866-1b3fde863816\") " Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.211002 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.211336 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:55.711318208 +0000 UTC m=+233.608692818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.211625 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12aae035-1b1b-4e29-8866-1b3fde863816-client-ca" (OuterVolumeSpecName: "client-ca") pod "12aae035-1b1b-4e29-8866-1b3fde863816" (UID: "12aae035-1b1b-4e29-8866-1b3fde863816"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.212365 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12aae035-1b1b-4e29-8866-1b3fde863816-config" (OuterVolumeSpecName: "config") pod "12aae035-1b1b-4e29-8866-1b3fde863816" (UID: "12aae035-1b1b-4e29-8866-1b3fde863816"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.227672 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12aae035-1b1b-4e29-8866-1b3fde863816-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "12aae035-1b1b-4e29-8866-1b3fde863816" (UID: "12aae035-1b1b-4e29-8866-1b3fde863816"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.227780 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12aae035-1b1b-4e29-8866-1b3fde863816-kube-api-access-nn8ff" (OuterVolumeSpecName: "kube-api-access-nn8ff") pod "12aae035-1b1b-4e29-8866-1b3fde863816" (UID: "12aae035-1b1b-4e29-8866-1b3fde863816"). InnerVolumeSpecName "kube-api-access-nn8ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.312587 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c616674-2247-42dd-9079-0ff8906d30da-config\") pod \"route-controller-manager-58ff7bcd8-w5scb\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.312678 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.312728 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c616674-2247-42dd-9079-0ff8906d30da-serving-cert\") pod \"route-controller-manager-58ff7bcd8-w5scb\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.312778 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c616674-2247-42dd-9079-0ff8906d30da-client-ca\") pod \"route-controller-manager-58ff7bcd8-w5scb\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.312829 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rghv4\" (UniqueName: \"kubernetes.io/projected/0c616674-2247-42dd-9079-0ff8906d30da-kube-api-access-rghv4\") pod \"route-controller-manager-58ff7bcd8-w5scb\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.312958 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12aae035-1b1b-4e29-8866-1b3fde863816-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.312973 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12aae035-1b1b-4e29-8866-1b3fde863816-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.312982 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12aae035-1b1b-4e29-8866-1b3fde863816-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.312991 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn8ff\" (UniqueName: \"kubernetes.io/projected/12aae035-1b1b-4e29-8866-1b3fde863816-kube-api-access-nn8ff\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.313210 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:55.813197738 +0000 UTC m=+233.710572348 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.413664 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.413966 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c616674-2247-42dd-9079-0ff8906d30da-serving-cert\") pod \"route-controller-manager-58ff7bcd8-w5scb\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.414010 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c616674-2247-42dd-9079-0ff8906d30da-client-ca\") pod \"route-controller-manager-58ff7bcd8-w5scb\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.414030 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rghv4\" (UniqueName: \"kubernetes.io/projected/0c616674-2247-42dd-9079-0ff8906d30da-kube-api-access-rghv4\") pod \"route-controller-manager-58ff7bcd8-w5scb\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.414071 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c616674-2247-42dd-9079-0ff8906d30da-config\") pod \"route-controller-manager-58ff7bcd8-w5scb\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.415298 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c616674-2247-42dd-9079-0ff8906d30da-config\") pod \"route-controller-manager-58ff7bcd8-w5scb\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.415374 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:55.915361397 +0000 UTC m=+233.812736007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.416917 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c616674-2247-42dd-9079-0ff8906d30da-client-ca\") pod \"route-controller-manager-58ff7bcd8-w5scb\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.426665 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c616674-2247-42dd-9079-0ff8906d30da-serving-cert\") pod \"route-controller-manager-58ff7bcd8-w5scb\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.476090 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.476316 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.478139 4983 patch_prober.go:28] interesting pod/console-f9d7485db-65l25 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.478295 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-65l25" podUID="6375157a-012b-441e-b97b-39d6830cef68" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.483321 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rghv4\" (UniqueName: \"kubernetes.io/projected/0c616674-2247-42dd-9079-0ff8906d30da-kube-api-access-rghv4\") pod \"route-controller-manager-58ff7bcd8-w5scb\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.514757 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.515097 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.015084423 +0000 UTC m=+233.912459033 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.604746 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8fpkl"] Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.605647 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.607295 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.615449 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.615539 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.11551669 +0000 UTC m=+234.012891310 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.615753 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.616124 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.116105968 +0000 UTC m=+234.013480578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.616136 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8fpkl"] Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.694802 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kzks7" event={"ID":"799072d7-3ce6-4742-be4c-065ecea7e24c","Type":"ContainerStarted","Data":"281559d8bd040460748ef989f60df59bff229ffdf02aa2cf03a61ee43d30b47b"} Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.695959 4983 generic.go:334] "Generic (PLEG): container finished" podID="12aae035-1b1b-4e29-8866-1b3fde863816" containerID="e285577cdd159cab74668fd228651187aec0ba78385e0118987b2db96c0ff114" exitCode=0 Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.696337 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" event={"ID":"12aae035-1b1b-4e29-8866-1b3fde863816","Type":"ContainerDied","Data":"e285577cdd159cab74668fd228651187aec0ba78385e0118987b2db96c0ff114"} Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.696401 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" event={"ID":"12aae035-1b1b-4e29-8866-1b3fde863816","Type":"ContainerDied","Data":"25eb8c6043db7ff2bff998648ec50b69c15942f4e0ad1fb2bdd0cc82e67399fc"} Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.696421 4983 scope.go:117] "RemoveContainer" containerID="e285577cdd159cab74668fd228651187aec0ba78385e0118987b2db96c0ff114" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.696604 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.702674 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.703352 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.713298 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.713402 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.713309 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.717208 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.717399 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.21737779 +0000 UTC m=+234.114752410 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.717440 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.717493 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe54f632-0209-4996-979c-4392b8f7d037-utilities\") pod \"community-operators-8fpkl\" (UID: \"fe54f632-0209-4996-979c-4392b8f7d037\") " pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.717533 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w8dv\" (UniqueName: \"kubernetes.io/projected/fe54f632-0209-4996-979c-4392b8f7d037-kube-api-access-9w8dv\") pod \"community-operators-8fpkl\" (UID: \"fe54f632-0209-4996-979c-4392b8f7d037\") " pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.717592 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe54f632-0209-4996-979c-4392b8f7d037-catalog-content\") pod \"community-operators-8fpkl\" (UID: \"fe54f632-0209-4996-979c-4392b8f7d037\") " pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.717904 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.217890035 +0000 UTC m=+234.115264645 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.735779 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-slqc5" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.774611 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.810109 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wnc92"] Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.811275 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.816173 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.819719 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.819996 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe54f632-0209-4996-979c-4392b8f7d037-catalog-content\") pod \"community-operators-8fpkl\" (UID: \"fe54f632-0209-4996-979c-4392b8f7d037\") " pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.820177 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe54f632-0209-4996-979c-4392b8f7d037-utilities\") pod \"community-operators-8fpkl\" (UID: \"fe54f632-0209-4996-979c-4392b8f7d037\") " pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.820236 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2bab65a6-74ae-4323-a9a8-bc3da3097d6f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2bab65a6-74ae-4323-a9a8-bc3da3097d6f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.820260 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w8dv\" (UniqueName: \"kubernetes.io/projected/fe54f632-0209-4996-979c-4392b8f7d037-kube-api-access-9w8dv\") pod \"community-operators-8fpkl\" (UID: \"fe54f632-0209-4996-979c-4392b8f7d037\") " pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.820283 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2bab65a6-74ae-4323-a9a8-bc3da3097d6f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2bab65a6-74ae-4323-a9a8-bc3da3097d6f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.820370 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.320356273 +0000 UTC m=+234.217730883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.821541 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe54f632-0209-4996-979c-4392b8f7d037-catalog-content\") pod \"community-operators-8fpkl\" (UID: \"fe54f632-0209-4996-979c-4392b8f7d037\") " pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.831695 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wnc92"] Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.832525 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe54f632-0209-4996-979c-4392b8f7d037-utilities\") pod \"community-operators-8fpkl\" (UID: \"fe54f632-0209-4996-979c-4392b8f7d037\") " pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.857822 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.860269 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4"] Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.863588 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g7fq4" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.865522 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84b4fdf79c-z4cv4"] Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.880601 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w8dv\" (UniqueName: \"kubernetes.io/projected/fe54f632-0209-4996-979c-4392b8f7d037-kube-api-access-9w8dv\") pod \"community-operators-8fpkl\" (UID: \"fe54f632-0209-4996-979c-4392b8f7d037\") " pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.922711 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.922946 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.923001 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a81a1b18-1976-45f7-97f5-4dcd7f745af3-utilities\") pod \"certified-operators-wnc92\" (UID: \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\") " pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.923021 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a81a1b18-1976-45f7-97f5-4dcd7f745af3-catalog-content\") pod \"certified-operators-wnc92\" (UID: \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\") " pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.923042 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2bab65a6-74ae-4323-a9a8-bc3da3097d6f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2bab65a6-74ae-4323-a9a8-bc3da3097d6f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.923062 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5znx9\" (UniqueName: \"kubernetes.io/projected/a81a1b18-1976-45f7-97f5-4dcd7f745af3-kube-api-access-5znx9\") pod \"certified-operators-wnc92\" (UID: \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\") " pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.923111 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2bab65a6-74ae-4323-a9a8-bc3da3097d6f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2bab65a6-74ae-4323-a9a8-bc3da3097d6f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.923300 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2bab65a6-74ae-4323-a9a8-bc3da3097d6f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2bab65a6-74ae-4323-a9a8-bc3da3097d6f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 00:11:55 crc kubenswrapper[4983]: E0318 00:11:55.924383 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.424369276 +0000 UTC m=+234.321743886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:55 crc kubenswrapper[4983]: I0318 00:11:55.958672 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2bab65a6-74ae-4323-a9a8-bc3da3097d6f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2bab65a6-74ae-4323-a9a8-bc3da3097d6f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.022032 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tdsf7"] Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.023003 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.026385 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.026719 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a81a1b18-1976-45f7-97f5-4dcd7f745af3-catalog-content\") pod \"certified-operators-wnc92\" (UID: \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\") " pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.026774 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a81a1b18-1976-45f7-97f5-4dcd7f745af3-utilities\") pod \"certified-operators-wnc92\" (UID: \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\") " pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.026800 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5znx9\" (UniqueName: \"kubernetes.io/projected/a81a1b18-1976-45f7-97f5-4dcd7f745af3-kube-api-access-5znx9\") pod \"certified-operators-wnc92\" (UID: \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\") " pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.027167 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.527150823 +0000 UTC m=+234.424525433 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.027503 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a81a1b18-1976-45f7-97f5-4dcd7f745af3-catalog-content\") pod \"certified-operators-wnc92\" (UID: \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\") " pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.027807 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a81a1b18-1976-45f7-97f5-4dcd7f745af3-utilities\") pod \"certified-operators-wnc92\" (UID: \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\") " pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.038382 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tdsf7"] Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.064691 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5znx9\" (UniqueName: \"kubernetes.io/projected/a81a1b18-1976-45f7-97f5-4dcd7f745af3-kube-api-access-5znx9\") pod \"certified-operators-wnc92\" (UID: \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\") " pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.103732 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.127563 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz95g\" (UniqueName: \"kubernetes.io/projected/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-kube-api-access-mz95g\") pod \"community-operators-tdsf7\" (UID: \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\") " pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.127615 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.127657 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-catalog-content\") pod \"community-operators-tdsf7\" (UID: \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\") " pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.127691 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-utilities\") pod \"community-operators-tdsf7\" (UID: \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\") " pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.128024 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.628012453 +0000 UTC m=+234.525387053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.134924 4983 patch_prober.go:28] interesting pod/downloads-7954f5f757-87bh5 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.134976 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-87bh5" podUID="eb9a12c3-1119-4b1e-8d32-0a9696a00392" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.134989 4983 patch_prober.go:28] interesting pod/downloads-7954f5f757-87bh5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.135043 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-87bh5" podUID="eb9a12c3-1119-4b1e-8d32-0a9696a00392" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.158706 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.176756 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.191125 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.194317 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:11:56 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:11:56 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:11:56 crc kubenswrapper[4983]: healthz check failed Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.194375 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.210128 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kxm7g"] Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.211317 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.228134 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.228314 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.728289466 +0000 UTC m=+234.625664086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.228377 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-catalog-content\") pod \"community-operators-tdsf7\" (UID: \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\") " pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.228421 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-utilities\") pod \"community-operators-tdsf7\" (UID: \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\") " pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.228517 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz95g\" (UniqueName: \"kubernetes.io/projected/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-kube-api-access-mz95g\") pod \"community-operators-tdsf7\" (UID: \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\") " pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.228573 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.229573 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.729562904 +0000 UTC m=+234.626937524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.230081 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-catalog-content\") pod \"community-operators-tdsf7\" (UID: \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\") " pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.230101 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-utilities\") pod \"community-operators-tdsf7\" (UID: \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\") " pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.239698 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kxm7g"] Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.251726 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz95g\" (UniqueName: \"kubernetes.io/projected/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-kube-api-access-mz95g\") pod \"community-operators-tdsf7\" (UID: \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\") " pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.255569 4983 patch_prober.go:28] interesting pod/apiserver-76f77b778f-h74lp container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 18 00:11:56 crc kubenswrapper[4983]: [+]log ok Mar 18 00:11:56 crc kubenswrapper[4983]: [+]etcd ok Mar 18 00:11:56 crc kubenswrapper[4983]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 18 00:11:56 crc kubenswrapper[4983]: [+]poststarthook/generic-apiserver-start-informers ok Mar 18 00:11:56 crc kubenswrapper[4983]: [+]poststarthook/max-in-flight-filter ok Mar 18 00:11:56 crc kubenswrapper[4983]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 18 00:11:56 crc kubenswrapper[4983]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 18 00:11:56 crc kubenswrapper[4983]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 18 00:11:56 crc kubenswrapper[4983]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Mar 18 00:11:56 crc kubenswrapper[4983]: [+]poststarthook/project.openshift.io-projectcache ok Mar 18 00:11:56 crc kubenswrapper[4983]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 18 00:11:56 crc kubenswrapper[4983]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Mar 18 00:11:56 crc kubenswrapper[4983]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 18 00:11:56 crc kubenswrapper[4983]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 18 00:11:56 crc kubenswrapper[4983]: livez check failed Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.255642 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-h74lp" podUID="53bc5818-4983-4f5f-8a3c-edc02aff5325" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.313678 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.314572 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.317056 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.319063 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.319298 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.330156 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.330352 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.830313461 +0000 UTC m=+234.727688071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.330404 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvvt9\" (UniqueName: \"kubernetes.io/projected/41c73828-774d-4440-a31d-c7cca08b063e-kube-api-access-bvvt9\") pod \"certified-operators-kxm7g\" (UID: \"41c73828-774d-4440-a31d-c7cca08b063e\") " pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.330489 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.330550 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c73828-774d-4440-a31d-c7cca08b063e-catalog-content\") pod \"certified-operators-kxm7g\" (UID: \"41c73828-774d-4440-a31d-c7cca08b063e\") " pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.330583 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c73828-774d-4440-a31d-c7cca08b063e-utilities\") pod \"certified-operators-kxm7g\" (UID: \"41c73828-774d-4440-a31d-c7cca08b063e\") " pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.330986 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.83097057 +0000 UTC m=+234.728345180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.342736 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.431669 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.431849 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.93181338 +0000 UTC m=+234.829188000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.431992 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.432041 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvvt9\" (UniqueName: \"kubernetes.io/projected/41c73828-774d-4440-a31d-c7cca08b063e-kube-api-access-bvvt9\") pod \"certified-operators-kxm7g\" (UID: \"41c73828-774d-4440-a31d-c7cca08b063e\") " pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.432076 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.432139 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.432329 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c73828-774d-4440-a31d-c7cca08b063e-catalog-content\") pod \"certified-operators-kxm7g\" (UID: \"41c73828-774d-4440-a31d-c7cca08b063e\") " pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.432354 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c73828-774d-4440-a31d-c7cca08b063e-utilities\") pod \"certified-operators-kxm7g\" (UID: \"41c73828-774d-4440-a31d-c7cca08b063e\") " pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.432575 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:56.932550982 +0000 UTC m=+234.829925672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.432703 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c73828-774d-4440-a31d-c7cca08b063e-utilities\") pod \"certified-operators-kxm7g\" (UID: \"41c73828-774d-4440-a31d-c7cca08b063e\") " pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.432868 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c73828-774d-4440-a31d-c7cca08b063e-catalog-content\") pod \"certified-operators-kxm7g\" (UID: \"41c73828-774d-4440-a31d-c7cca08b063e\") " pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.447184 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvvt9\" (UniqueName: \"kubernetes.io/projected/41c73828-774d-4440-a31d-c7cca08b063e-kube-api-access-bvvt9\") pod \"certified-operators-kxm7g\" (UID: \"41c73828-774d-4440-a31d-c7cca08b063e\") " pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.502496 4983 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.516399 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-wsk9g" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.537260 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.537533 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.537613 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.537905 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:57.037887175 +0000 UTC m=+234.935261785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.537952 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.538812 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.543222 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8qf4k" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.558349 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.590733 4983 ???:1] "http: TLS handshake error from 192.168.126.11:53352: no serving certificate available for the kubelet" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.627772 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.639959 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.640889 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:57.140877108 +0000 UTC m=+235.038251718 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.741622 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.741966 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:57.241924124 +0000 UTC m=+235.139298734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.843270 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.844184 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:57.344141394 +0000 UTC m=+235.241516064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.852723 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12aae035-1b1b-4e29-8866-1b3fde863816" path="/var/lib/kubelet/pods/12aae035-1b1b-4e29-8866-1b3fde863816/volumes" Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.945274 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.945300 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:57.445277632 +0000 UTC m=+235.342652242 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:56 crc kubenswrapper[4983]: I0318 00:11:56.945634 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:56 crc kubenswrapper[4983]: E0318 00:11:56.946029 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:57.446012374 +0000 UTC m=+235.343386984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.047143 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:57 crc kubenswrapper[4983]: E0318 00:11:57.047275 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 00:11:57.547250665 +0000 UTC m=+235.444625275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.047389 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:57 crc kubenswrapper[4983]: E0318 00:11:57.047776 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 00:11:57.547765461 +0000 UTC m=+235.445140071 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qkkt2" (UID: "297c1426-c499-449c-ab44-e1f898262839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.066142 4983 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-18T00:11:56.5025213Z","Handler":null,"Name":""} Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.075078 4983 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.075120 4983 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.148047 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.153828 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.196935 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:11:57 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:11:57 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:11:57 crc kubenswrapper[4983]: healthz check failed Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.197003 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.251466 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.254654 4983 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.255156 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.301983 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qkkt2\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.554631 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.709246 4983 generic.go:334] "Generic (PLEG): container finished" podID="602912b1-d12a-406a-ba6e-e0d94dc62038" containerID="21093baebef5e62832dacba00ecd7e93e6fd4fd43659782d70b3341bf1eb3209" exitCode=0 Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.709290 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" event={"ID":"602912b1-d12a-406a-ba6e-e0d94dc62038","Type":"ContainerDied","Data":"21093baebef5e62832dacba00ecd7e93e6fd4fd43659782d70b3341bf1eb3209"} Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.807129 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lqjwd"] Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.808865 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.812499 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.813159 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lqjwd"] Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.860011 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5099d24e-b4a3-4990-bc90-980e0188ef3e-catalog-content\") pod \"redhat-marketplace-lqjwd\" (UID: \"5099d24e-b4a3-4990-bc90-980e0188ef3e\") " pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.860068 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5099d24e-b4a3-4990-bc90-980e0188ef3e-utilities\") pod \"redhat-marketplace-lqjwd\" (UID: \"5099d24e-b4a3-4990-bc90-980e0188ef3e\") " pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.860123 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66hvg\" (UniqueName: \"kubernetes.io/projected/5099d24e-b4a3-4990-bc90-980e0188ef3e-kube-api-access-66hvg\") pod \"redhat-marketplace-lqjwd\" (UID: \"5099d24e-b4a3-4990-bc90-980e0188ef3e\") " pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.961898 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66hvg\" (UniqueName: \"kubernetes.io/projected/5099d24e-b4a3-4990-bc90-980e0188ef3e-kube-api-access-66hvg\") pod \"redhat-marketplace-lqjwd\" (UID: \"5099d24e-b4a3-4990-bc90-980e0188ef3e\") " pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.962010 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5099d24e-b4a3-4990-bc90-980e0188ef3e-catalog-content\") pod \"redhat-marketplace-lqjwd\" (UID: \"5099d24e-b4a3-4990-bc90-980e0188ef3e\") " pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.962030 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5099d24e-b4a3-4990-bc90-980e0188ef3e-utilities\") pod \"redhat-marketplace-lqjwd\" (UID: \"5099d24e-b4a3-4990-bc90-980e0188ef3e\") " pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.962474 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5099d24e-b4a3-4990-bc90-980e0188ef3e-utilities\") pod \"redhat-marketplace-lqjwd\" (UID: \"5099d24e-b4a3-4990-bc90-980e0188ef3e\") " pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.962521 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5099d24e-b4a3-4990-bc90-980e0188ef3e-catalog-content\") pod \"redhat-marketplace-lqjwd\" (UID: \"5099d24e-b4a3-4990-bc90-980e0188ef3e\") " pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:11:57 crc kubenswrapper[4983]: I0318 00:11:57.982674 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66hvg\" (UniqueName: \"kubernetes.io/projected/5099d24e-b4a3-4990-bc90-980e0188ef3e-kube-api-access-66hvg\") pod \"redhat-marketplace-lqjwd\" (UID: \"5099d24e-b4a3-4990-bc90-980e0188ef3e\") " pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.135425 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.192522 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:11:58 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:11:58 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:11:58 crc kubenswrapper[4983]: healthz check failed Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.192598 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.203017 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fs978"] Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.204548 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.215398 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fs978"] Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.265537 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089eb8e9-84d9-4320-97ec-9d868b858d3a-utilities\") pod \"redhat-marketplace-fs978\" (UID: \"089eb8e9-84d9-4320-97ec-9d868b858d3a\") " pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.265610 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssbd7\" (UniqueName: \"kubernetes.io/projected/089eb8e9-84d9-4320-97ec-9d868b858d3a-kube-api-access-ssbd7\") pod \"redhat-marketplace-fs978\" (UID: \"089eb8e9-84d9-4320-97ec-9d868b858d3a\") " pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.265672 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089eb8e9-84d9-4320-97ec-9d868b858d3a-catalog-content\") pod \"redhat-marketplace-fs978\" (UID: \"089eb8e9-84d9-4320-97ec-9d868b858d3a\") " pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.325128 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-k424z" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.366560 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089eb8e9-84d9-4320-97ec-9d868b858d3a-catalog-content\") pod \"redhat-marketplace-fs978\" (UID: \"089eb8e9-84d9-4320-97ec-9d868b858d3a\") " pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.366609 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089eb8e9-84d9-4320-97ec-9d868b858d3a-utilities\") pod \"redhat-marketplace-fs978\" (UID: \"089eb8e9-84d9-4320-97ec-9d868b858d3a\") " pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.366663 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssbd7\" (UniqueName: \"kubernetes.io/projected/089eb8e9-84d9-4320-97ec-9d868b858d3a-kube-api-access-ssbd7\") pod \"redhat-marketplace-fs978\" (UID: \"089eb8e9-84d9-4320-97ec-9d868b858d3a\") " pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.367577 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089eb8e9-84d9-4320-97ec-9d868b858d3a-catalog-content\") pod \"redhat-marketplace-fs978\" (UID: \"089eb8e9-84d9-4320-97ec-9d868b858d3a\") " pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.367625 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089eb8e9-84d9-4320-97ec-9d868b858d3a-utilities\") pod \"redhat-marketplace-fs978\" (UID: \"089eb8e9-84d9-4320-97ec-9d868b858d3a\") " pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.388721 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssbd7\" (UniqueName: \"kubernetes.io/projected/089eb8e9-84d9-4320-97ec-9d868b858d3a-kube-api-access-ssbd7\") pod \"redhat-marketplace-fs978\" (UID: \"089eb8e9-84d9-4320-97ec-9d868b858d3a\") " pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.529315 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.780983 4983 scope.go:117] "RemoveContainer" containerID="e285577cdd159cab74668fd228651187aec0ba78385e0118987b2db96c0ff114" Mar 18 00:11:58 crc kubenswrapper[4983]: E0318 00:11:58.781785 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e285577cdd159cab74668fd228651187aec0ba78385e0118987b2db96c0ff114\": container with ID starting with e285577cdd159cab74668fd228651187aec0ba78385e0118987b2db96c0ff114 not found: ID does not exist" containerID="e285577cdd159cab74668fd228651187aec0ba78385e0118987b2db96c0ff114" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.781847 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e285577cdd159cab74668fd228651187aec0ba78385e0118987b2db96c0ff114"} err="failed to get container status \"e285577cdd159cab74668fd228651187aec0ba78385e0118987b2db96c0ff114\": rpc error: code = NotFound desc = could not find container \"e285577cdd159cab74668fd228651187aec0ba78385e0118987b2db96c0ff114\": container with ID starting with e285577cdd159cab74668fd228651187aec0ba78385e0118987b2db96c0ff114 not found: ID does not exist" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.805434 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-78tfj"] Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.807187 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.809493 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.814907 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-78tfj"] Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.864468 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.874699 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75077192-24a6-4efb-8484-bea2db4a4812-utilities\") pod \"redhat-operators-78tfj\" (UID: \"75077192-24a6-4efb-8484-bea2db4a4812\") " pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.874752 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75077192-24a6-4efb-8484-bea2db4a4812-catalog-content\") pod \"redhat-operators-78tfj\" (UID: \"75077192-24a6-4efb-8484-bea2db4a4812\") " pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.874847 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xjtj\" (UniqueName: \"kubernetes.io/projected/75077192-24a6-4efb-8484-bea2db4a4812-kube-api-access-7xjtj\") pod \"redhat-operators-78tfj\" (UID: \"75077192-24a6-4efb-8484-bea2db4a4812\") " pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.975877 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xjtj\" (UniqueName: \"kubernetes.io/projected/75077192-24a6-4efb-8484-bea2db4a4812-kube-api-access-7xjtj\") pod \"redhat-operators-78tfj\" (UID: \"75077192-24a6-4efb-8484-bea2db4a4812\") " pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.977615 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75077192-24a6-4efb-8484-bea2db4a4812-utilities\") pod \"redhat-operators-78tfj\" (UID: \"75077192-24a6-4efb-8484-bea2db4a4812\") " pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.978362 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75077192-24a6-4efb-8484-bea2db4a4812-catalog-content\") pod \"redhat-operators-78tfj\" (UID: \"75077192-24a6-4efb-8484-bea2db4a4812\") " pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.978290 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75077192-24a6-4efb-8484-bea2db4a4812-utilities\") pod \"redhat-operators-78tfj\" (UID: \"75077192-24a6-4efb-8484-bea2db4a4812\") " pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:11:58 crc kubenswrapper[4983]: I0318 00:11:58.979017 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75077192-24a6-4efb-8484-bea2db4a4812-catalog-content\") pod \"redhat-operators-78tfj\" (UID: \"75077192-24a6-4efb-8484-bea2db4a4812\") " pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.003148 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xjtj\" (UniqueName: \"kubernetes.io/projected/75077192-24a6-4efb-8484-bea2db4a4812-kube-api-access-7xjtj\") pod \"redhat-operators-78tfj\" (UID: \"75077192-24a6-4efb-8484-bea2db4a4812\") " pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.149816 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.195338 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:11:59 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:11:59 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:11:59 crc kubenswrapper[4983]: healthz check failed Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.197151 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.208103 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t5mgj"] Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.209114 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.220556 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t5mgj"] Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.283360 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-catalog-content\") pod \"redhat-operators-t5mgj\" (UID: \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\") " pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.283444 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-utilities\") pod \"redhat-operators-t5mgj\" (UID: \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\") " pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.283510 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72btz\" (UniqueName: \"kubernetes.io/projected/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-kube-api-access-72btz\") pod \"redhat-operators-t5mgj\" (UID: \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\") " pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.344076 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.384211 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/602912b1-d12a-406a-ba6e-e0d94dc62038-secret-volume\") pod \"602912b1-d12a-406a-ba6e-e0d94dc62038\" (UID: \"602912b1-d12a-406a-ba6e-e0d94dc62038\") " Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.384284 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/602912b1-d12a-406a-ba6e-e0d94dc62038-config-volume\") pod \"602912b1-d12a-406a-ba6e-e0d94dc62038\" (UID: \"602912b1-d12a-406a-ba6e-e0d94dc62038\") " Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.384321 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r824j\" (UniqueName: \"kubernetes.io/projected/602912b1-d12a-406a-ba6e-e0d94dc62038-kube-api-access-r824j\") pod \"602912b1-d12a-406a-ba6e-e0d94dc62038\" (UID: \"602912b1-d12a-406a-ba6e-e0d94dc62038\") " Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.384505 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72btz\" (UniqueName: \"kubernetes.io/projected/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-kube-api-access-72btz\") pod \"redhat-operators-t5mgj\" (UID: \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\") " pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.384592 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-catalog-content\") pod \"redhat-operators-t5mgj\" (UID: \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\") " pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.384612 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-utilities\") pod \"redhat-operators-t5mgj\" (UID: \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\") " pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.386676 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/602912b1-d12a-406a-ba6e-e0d94dc62038-config-volume" (OuterVolumeSpecName: "config-volume") pod "602912b1-d12a-406a-ba6e-e0d94dc62038" (UID: "602912b1-d12a-406a-ba6e-e0d94dc62038"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.386841 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-catalog-content\") pod \"redhat-operators-t5mgj\" (UID: \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\") " pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.387234 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-utilities\") pod \"redhat-operators-t5mgj\" (UID: \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\") " pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.388756 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/602912b1-d12a-406a-ba6e-e0d94dc62038-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "602912b1-d12a-406a-ba6e-e0d94dc62038" (UID: "602912b1-d12a-406a-ba6e-e0d94dc62038"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.388838 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/602912b1-d12a-406a-ba6e-e0d94dc62038-kube-api-access-r824j" (OuterVolumeSpecName: "kube-api-access-r824j") pod "602912b1-d12a-406a-ba6e-e0d94dc62038" (UID: "602912b1-d12a-406a-ba6e-e0d94dc62038"). InnerVolumeSpecName "kube-api-access-r824j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.403498 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72btz\" (UniqueName: \"kubernetes.io/projected/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-kube-api-access-72btz\") pod \"redhat-operators-t5mgj\" (UID: \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\") " pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.439873 4983 ???:1] "http: TLS handshake error from 192.168.126.11:53356: no serving certificate available for the kubelet" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.486033 4983 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/602912b1-d12a-406a-ba6e-e0d94dc62038-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.486069 4983 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/602912b1-d12a-406a-ba6e-e0d94dc62038-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.486080 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r824j\" (UniqueName: \"kubernetes.io/projected/602912b1-d12a-406a-ba6e-e0d94dc62038-kube-api-access-r824j\") on node \"crc\" DevicePath \"\"" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.540013 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.721545 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" event={"ID":"602912b1-d12a-406a-ba6e-e0d94dc62038","Type":"ContainerDied","Data":"04802dd1029d53bf5e83029b4a1b711560104f8a7f71db882592357af4d18264"} Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.721581 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04802dd1029d53bf5e83029b4a1b711560104f8a7f71db882592357af4d18264" Mar 18 00:11:59 crc kubenswrapper[4983]: I0318 00:11:59.721628 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563200-xtxcb" Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.134097 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563212-g7qg2"] Mar 18 00:12:00 crc kubenswrapper[4983]: E0318 00:12:00.134319 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="602912b1-d12a-406a-ba6e-e0d94dc62038" containerName="collect-profiles" Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.134331 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="602912b1-d12a-406a-ba6e-e0d94dc62038" containerName="collect-profiles" Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.134441 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="602912b1-d12a-406a-ba6e-e0d94dc62038" containerName="collect-profiles" Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.134844 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563212-g7qg2" Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.139687 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.145634 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563212-g7qg2"] Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.194294 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:12:00 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:12:00 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:12:00 crc kubenswrapper[4983]: healthz check failed Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.194369 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.195578 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btznr\" (UniqueName: \"kubernetes.io/projected/55e41d80-a523-4a82-8d42-c01daa50d832-kube-api-access-btznr\") pod \"auto-csr-approver-29563212-g7qg2\" (UID: \"55e41d80-a523-4a82-8d42-c01daa50d832\") " pod="openshift-infra/auto-csr-approver-29563212-g7qg2" Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.296780 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btznr\" (UniqueName: \"kubernetes.io/projected/55e41d80-a523-4a82-8d42-c01daa50d832-kube-api-access-btznr\") pod \"auto-csr-approver-29563212-g7qg2\" (UID: \"55e41d80-a523-4a82-8d42-c01daa50d832\") " pod="openshift-infra/auto-csr-approver-29563212-g7qg2" Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.323639 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btznr\" (UniqueName: \"kubernetes.io/projected/55e41d80-a523-4a82-8d42-c01daa50d832-kube-api-access-btznr\") pod \"auto-csr-approver-29563212-g7qg2\" (UID: \"55e41d80-a523-4a82-8d42-c01daa50d832\") " pod="openshift-infra/auto-csr-approver-29563212-g7qg2" Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.485321 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563212-g7qg2" Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.595704 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:12:00 crc kubenswrapper[4983]: I0318 00:12:00.603392 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-h74lp" Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.197327 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:12:01 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:12:01 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:12:01 crc kubenswrapper[4983]: healthz check failed Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.197577 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:12:01 crc kubenswrapper[4983]: W0318 00:12:01.226142 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe54f632_0209_4996_979c_4392b8f7d037.slice/crio-091dd67e5a9c45c4d7cacd0e54c868bd4c82fde5b4629f06ef7156f6373ec434 WatchSource:0}: Error finding container 091dd67e5a9c45c4d7cacd0e54c868bd4c82fde5b4629f06ef7156f6373ec434: Status 404 returned error can't find the container with id 091dd67e5a9c45c4d7cacd0e54c868bd4c82fde5b4629f06ef7156f6373ec434 Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.226716 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8fpkl"] Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.488342 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lqjwd"] Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.500102 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.506778 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kxm7g"] Mar 18 00:12:01 crc kubenswrapper[4983]: W0318 00:12:01.515442 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2bab65a6_74ae_4323_a9a8_bc3da3097d6f.slice/crio-1e0ad35f2ad6c446e25d7a5af4a6c7038030b2c1b3f1073d325f758d94382e7b WatchSource:0}: Error finding container 1e0ad35f2ad6c446e25d7a5af4a6c7038030b2c1b3f1073d325f758d94382e7b: Status 404 returned error can't find the container with id 1e0ad35f2ad6c446e25d7a5af4a6c7038030b2c1b3f1073d325f758d94382e7b Mar 18 00:12:01 crc kubenswrapper[4983]: W0318 00:12:01.519335 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5099d24e_b4a3_4990_bc90_980e0188ef3e.slice/crio-b30a9dda087848793ea13b5099ed5bdbe2deb683aa09faef3825ec8196490093 WatchSource:0}: Error finding container b30a9dda087848793ea13b5099ed5bdbe2deb683aa09faef3825ec8196490093: Status 404 returned error can't find the container with id b30a9dda087848793ea13b5099ed5bdbe2deb683aa09faef3825ec8196490093 Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.527912 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fs978"] Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.532340 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563212-g7qg2"] Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.551652 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.584513 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wnc92"] Mar 18 00:12:01 crc kubenswrapper[4983]: W0318 00:12:01.612797 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda81a1b18_1976_45f7_97f5_4dcd7f745af3.slice/crio-e5b75985bcc2c379b15b8dc0572e90ff16ffa2dfa05aaa6ee8e21dc54fb68d45 WatchSource:0}: Error finding container e5b75985bcc2c379b15b8dc0572e90ff16ffa2dfa05aaa6ee8e21dc54fb68d45: Status 404 returned error can't find the container with id e5b75985bcc2c379b15b8dc0572e90ff16ffa2dfa05aaa6ee8e21dc54fb68d45 Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.614445 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb"] Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.618229 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tdsf7"] Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.624911 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t5mgj"] Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.644282 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qkkt2"] Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.646058 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-78tfj"] Mar 18 00:12:01 crc kubenswrapper[4983]: W0318 00:12:01.669805 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c616674_2247_42dd_9079_0ff8906d30da.slice/crio-f387cfa53894e2bf23dcc323f4ca61dc029f8fa193ecedcf1ed6123d414bc8a9 WatchSource:0}: Error finding container f387cfa53894e2bf23dcc323f4ca61dc029f8fa193ecedcf1ed6123d414bc8a9: Status 404 returned error can't find the container with id f387cfa53894e2bf23dcc323f4ca61dc029f8fa193ecedcf1ed6123d414bc8a9 Mar 18 00:12:01 crc kubenswrapper[4983]: W0318 00:12:01.672419 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e139d3f_fd2e_4e39_96b4_eafd340c10a3.slice/crio-6b6fea9a4ccf1425c159036414b60551fc1946ace7ba4bc1f3ee2abfc8050b25 WatchSource:0}: Error finding container 6b6fea9a4ccf1425c159036414b60551fc1946ace7ba4bc1f3ee2abfc8050b25: Status 404 returned error can't find the container with id 6b6fea9a4ccf1425c159036414b60551fc1946ace7ba4bc1f3ee2abfc8050b25 Mar 18 00:12:01 crc kubenswrapper[4983]: W0318 00:12:01.675465 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75077192_24a6_4efb_8484_bea2db4a4812.slice/crio-86be496c37a20089eaba15e2afe206f29958e6e4e36a6be2b9d667e5fe025a9c WatchSource:0}: Error finding container 86be496c37a20089eaba15e2afe206f29958e6e4e36a6be2b9d667e5fe025a9c: Status 404 returned error can't find the container with id 86be496c37a20089eaba15e2afe206f29958e6e4e36a6be2b9d667e5fe025a9c Mar 18 00:12:01 crc kubenswrapper[4983]: W0318 00:12:01.678158 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod297c1426_c499_449c_ab44_e1f898262839.slice/crio-37776685707eb97ab07718f65b6b31581627b8cc12e80ea52a82dafa2914945b WatchSource:0}: Error finding container 37776685707eb97ab07718f65b6b31581627b8cc12e80ea52a82dafa2914945b: Status 404 returned error can't find the container with id 37776685707eb97ab07718f65b6b31581627b8cc12e80ea52a82dafa2914945b Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.721825 4983 csr.go:261] certificate signing request csr-dxs5g is approved, waiting to be issued Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.728542 4983 csr.go:257] certificate signing request csr-dxs5g is issued Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.746456 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" event={"ID":"297c1426-c499-449c-ab44-e1f898262839","Type":"ContainerStarted","Data":"37776685707eb97ab07718f65b6b31581627b8cc12e80ea52a82dafa2914945b"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.748129 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78tfj" event={"ID":"75077192-24a6-4efb-8484-bea2db4a4812","Type":"ContainerStarted","Data":"86be496c37a20089eaba15e2afe206f29958e6e4e36a6be2b9d667e5fe025a9c"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.752243 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kzks7" event={"ID":"799072d7-3ce6-4742-be4c-065ecea7e24c","Type":"ContainerStarted","Data":"0ff85708214a76a4f6653f8627ef484a76131250027c73db3c5ee9f23c53dfaf"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.752268 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kzks7" event={"ID":"799072d7-3ce6-4742-be4c-065ecea7e24c","Type":"ContainerStarted","Data":"513ad85aa508ebbc3fd26d6204af8d73c51de8b9e1fbf7720a366369a162edd9"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.756994 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fs978" event={"ID":"089eb8e9-84d9-4320-97ec-9d868b858d3a","Type":"ContainerStarted","Data":"ff9d1df6bac87226e656c720843582a61e3dc5b1f4be7103b31cb43709f2f298"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.759374 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdsf7" event={"ID":"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419","Type":"ContainerStarted","Data":"49c28dcce297ab899b65e2b0b8348b7f7d53f5c2ca574c1c69964ccd90e77b90"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.760367 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxm7g" event={"ID":"41c73828-774d-4440-a31d-c7cca08b063e","Type":"ContainerStarted","Data":"f22a7aa36053005ecbfcf5f5630ac8e00b25957cb8d9814ad3d44a8a65e8fc14"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.761445 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563212-g7qg2" event={"ID":"55e41d80-a523-4a82-8d42-c01daa50d832","Type":"ContainerStarted","Data":"c741f78e43e4310dede5dd1f652f3cd709e87461bc1bc84ce80accdb862f9529"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.762092 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef","Type":"ContainerStarted","Data":"2c6a6c9cb65162fc4c7289ba726f1fd801cd5808cec47702dff17e8d0676dc47"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.764761 4983 generic.go:334] "Generic (PLEG): container finished" podID="fe54f632-0209-4996-979c-4392b8f7d037" containerID="498ac344d5498ae2dea1c12999eae6fe8127ea6d3c5628439d9563533fcf9074" exitCode=0 Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.764819 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fpkl" event={"ID":"fe54f632-0209-4996-979c-4392b8f7d037","Type":"ContainerDied","Data":"498ac344d5498ae2dea1c12999eae6fe8127ea6d3c5628439d9563533fcf9074"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.764837 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fpkl" event={"ID":"fe54f632-0209-4996-979c-4392b8f7d037","Type":"ContainerStarted","Data":"091dd67e5a9c45c4d7cacd0e54c868bd4c82fde5b4629f06ef7156f6373ec434"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.766516 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" event={"ID":"0c616674-2247-42dd-9079-0ff8906d30da","Type":"ContainerStarted","Data":"f387cfa53894e2bf23dcc323f4ca61dc029f8fa193ecedcf1ed6123d414bc8a9"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.768155 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5mgj" event={"ID":"5e139d3f-fd2e-4e39-96b4-eafd340c10a3","Type":"ContainerStarted","Data":"6b6fea9a4ccf1425c159036414b60551fc1946ace7ba4bc1f3ee2abfc8050b25"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.769720 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lqjwd" event={"ID":"5099d24e-b4a3-4990-bc90-980e0188ef3e","Type":"ContainerStarted","Data":"b30a9dda087848793ea13b5099ed5bdbe2deb683aa09faef3825ec8196490093"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.771057 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563210-lptd5" event={"ID":"efeae5ef-dba6-4031-83be-8f493fe853b9","Type":"ContainerStarted","Data":"a7d0b17358ee5ef5868fa4c70dc42c060581f32d2f5170f73dc9e44fb25e5400"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.771221 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-kzks7" podStartSLOduration=18.771210168 podStartE2EDuration="18.771210168s" podCreationTimestamp="2026-03-18 00:11:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:12:01.768721434 +0000 UTC m=+239.666096054" watchObservedRunningTime="2026-03-18 00:12:01.771210168 +0000 UTC m=+239.668584778" Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.774073 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2bab65a6-74ae-4323-a9a8-bc3da3097d6f","Type":"ContainerStarted","Data":"1e0ad35f2ad6c446e25d7a5af4a6c7038030b2c1b3f1073d325f758d94382e7b"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.774685 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnc92" event={"ID":"a81a1b18-1976-45f7-97f5-4dcd7f745af3","Type":"ContainerStarted","Data":"e5b75985bcc2c379b15b8dc0572e90ff16ffa2dfa05aaa6ee8e21dc54fb68d45"} Mar 18 00:12:01 crc kubenswrapper[4983]: I0318 00:12:01.797119 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29563210-lptd5" podStartSLOduration=109.140496722 podStartE2EDuration="2m1.79710166s" podCreationTimestamp="2026-03-18 00:10:00 +0000 UTC" firstStartedPulling="2026-03-18 00:11:48.223229889 +0000 UTC m=+226.120604499" lastFinishedPulling="2026-03-18 00:12:00.879834827 +0000 UTC m=+238.777209437" observedRunningTime="2026-03-18 00:12:01.79643177 +0000 UTC m=+239.693806380" watchObservedRunningTime="2026-03-18 00:12:01.79710166 +0000 UTC m=+239.694476270" Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.195375 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:12:02 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:12:02 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:12:02 crc kubenswrapper[4983]: healthz check failed Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.195434 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.729863 4983 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-15 02:46:40.185206881 +0000 UTC Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.730245 4983 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 5810h34m37.454966154s for next certificate rotation Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.804295 4983 generic.go:334] "Generic (PLEG): container finished" podID="089eb8e9-84d9-4320-97ec-9d868b858d3a" containerID="72fe3e72ff617dd5535918aea93cd17cff83181b2b012c44a2e3d2511ab3189e" exitCode=0 Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.804385 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fs978" event={"ID":"089eb8e9-84d9-4320-97ec-9d868b858d3a","Type":"ContainerDied","Data":"72fe3e72ff617dd5535918aea93cd17cff83181b2b012c44a2e3d2511ab3189e"} Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.833685 4983 generic.go:334] "Generic (PLEG): container finished" podID="41c73828-774d-4440-a31d-c7cca08b063e" containerID="8f61cdd4663f4becd3fad3fa742dc3a077ac4cb060a0fd09c831e16cb38bbbbd" exitCode=0 Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.833769 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxm7g" event={"ID":"41c73828-774d-4440-a31d-c7cca08b063e","Type":"ContainerDied","Data":"8f61cdd4663f4becd3fad3fa742dc3a077ac4cb060a0fd09c831e16cb38bbbbd"} Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.841810 4983 generic.go:334] "Generic (PLEG): container finished" podID="b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" containerID="1bd334234e1433a96fc41dbf0a6d47cf493ed4dcb3d95118d38b638aee621620" exitCode=0 Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.841896 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdsf7" event={"ID":"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419","Type":"ContainerDied","Data":"1bd334234e1433a96fc41dbf0a6d47cf493ed4dcb3d95118d38b638aee621620"} Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.859018 4983 generic.go:334] "Generic (PLEG): container finished" podID="5e139d3f-fd2e-4e39-96b4-eafd340c10a3" containerID="453aca7b0fae9e0bc9b795bf327ca036152e6e08018a1b820f2bbc94b6f83332" exitCode=0 Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.867910 4983 generic.go:334] "Generic (PLEG): container finished" podID="efeae5ef-dba6-4031-83be-8f493fe853b9" containerID="a7d0b17358ee5ef5868fa4c70dc42c060581f32d2f5170f73dc9e44fb25e5400" exitCode=0 Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.869809 4983 generic.go:334] "Generic (PLEG): container finished" podID="2bab65a6-74ae-4323-a9a8-bc3da3097d6f" containerID="b20fdab19e57871a8f43f4458fb334831dfd78b03a0563e5a0bce1f6375db559" exitCode=0 Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.871991 4983 generic.go:334] "Generic (PLEG): container finished" podID="75077192-24a6-4efb-8484-bea2db4a4812" containerID="0c48bab5388a827e7984b67764d8bb7410e773eda353b791433bb37068deedc5" exitCode=0 Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.874353 4983 generic.go:334] "Generic (PLEG): container finished" podID="a81a1b18-1976-45f7-97f5-4dcd7f745af3" containerID="203afe93d4764619d6478248cfea556a15b41013072fbd9efe114ade133b9a42" exitCode=0 Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.879167 4983 generic.go:334] "Generic (PLEG): container finished" podID="5099d24e-b4a3-4990-bc90-980e0188ef3e" containerID="25781ee37f4c893c4ce77bc9b9f63be73b3ade83b0bae39e0f1ae4ad3f020339" exitCode=0 Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.894976 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5mgj" event={"ID":"5e139d3f-fd2e-4e39-96b4-eafd340c10a3","Type":"ContainerDied","Data":"453aca7b0fae9e0bc9b795bf327ca036152e6e08018a1b820f2bbc94b6f83332"} Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.895015 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.895028 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" event={"ID":"297c1426-c499-449c-ab44-e1f898262839","Type":"ContainerStarted","Data":"36ac6c57e9ecbf8bed1f3803e76cd664da2818fb62f8e12aef88b060b41af978"} Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.895037 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.895047 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef","Type":"ContainerStarted","Data":"7a90617b3b0d6771c977d6c48b45ea4ff4fc981c02f5e742dc62296448b86f5a"} Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.895062 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563210-lptd5" event={"ID":"efeae5ef-dba6-4031-83be-8f493fe853b9","Type":"ContainerDied","Data":"a7d0b17358ee5ef5868fa4c70dc42c060581f32d2f5170f73dc9e44fb25e5400"} Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.895072 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2bab65a6-74ae-4323-a9a8-bc3da3097d6f","Type":"ContainerDied","Data":"b20fdab19e57871a8f43f4458fb334831dfd78b03a0563e5a0bce1f6375db559"} Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.895083 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78tfj" event={"ID":"75077192-24a6-4efb-8484-bea2db4a4812","Type":"ContainerDied","Data":"0c48bab5388a827e7984b67764d8bb7410e773eda353b791433bb37068deedc5"} Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.895092 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnc92" event={"ID":"a81a1b18-1976-45f7-97f5-4dcd7f745af3","Type":"ContainerDied","Data":"203afe93d4764619d6478248cfea556a15b41013072fbd9efe114ade133b9a42"} Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.895101 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lqjwd" event={"ID":"5099d24e-b4a3-4990-bc90-980e0188ef3e","Type":"ContainerDied","Data":"25781ee37f4c893c4ce77bc9b9f63be73b3ade83b0bae39e0f1ae4ad3f020339"} Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.895113 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" event={"ID":"0c616674-2247-42dd-9079-0ff8906d30da","Type":"ContainerStarted","Data":"33795ce60bdb6d48e29737e87fc9580caacc10c25d0ff4334bd8828fdce2856e"} Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.903485 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:12:02 crc kubenswrapper[4983]: I0318 00:12:02.967704 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" podStartSLOduration=178.967691623 podStartE2EDuration="2m58.967691623s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:12:02.96723286 +0000 UTC m=+240.864607480" watchObservedRunningTime="2026-03-18 00:12:02.967691623 +0000 UTC m=+240.865066233" Mar 18 00:12:03 crc kubenswrapper[4983]: I0318 00:12:03.017163 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" podStartSLOduration=11.017146018 podStartE2EDuration="11.017146018s" podCreationTimestamp="2026-03-18 00:11:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:12:03.015775227 +0000 UTC m=+240.913149847" watchObservedRunningTime="2026-03-18 00:12:03.017146018 +0000 UTC m=+240.914520628" Mar 18 00:12:03 crc kubenswrapper[4983]: I0318 00:12:03.195677 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:12:03 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:12:03 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:12:03 crc kubenswrapper[4983]: healthz check failed Mar 18 00:12:03 crc kubenswrapper[4983]: I0318 00:12:03.195726 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:12:03 crc kubenswrapper[4983]: I0318 00:12:03.730895 4983 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-01 10:49:35.865981093 +0000 UTC Mar 18 00:12:03 crc kubenswrapper[4983]: I0318 00:12:03.730963 4983 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6202h37m32.135022373s for next certificate rotation Mar 18 00:12:03 crc kubenswrapper[4983]: I0318 00:12:03.905875 4983 generic.go:334] "Generic (PLEG): container finished" podID="55e41d80-a523-4a82-8d42-c01daa50d832" containerID="c680d35f60a7e80033026aef33a6e7b38bf5ef016a8e826b865081310759914d" exitCode=0 Mar 18 00:12:03 crc kubenswrapper[4983]: I0318 00:12:03.905947 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563212-g7qg2" event={"ID":"55e41d80-a523-4a82-8d42-c01daa50d832","Type":"ContainerDied","Data":"c680d35f60a7e80033026aef33a6e7b38bf5ef016a8e826b865081310759914d"} Mar 18 00:12:03 crc kubenswrapper[4983]: I0318 00:12:03.915394 4983 generic.go:334] "Generic (PLEG): container finished" podID="ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef" containerID="7a90617b3b0d6771c977d6c48b45ea4ff4fc981c02f5e742dc62296448b86f5a" exitCode=0 Mar 18 00:12:03 crc kubenswrapper[4983]: I0318 00:12:03.915667 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef","Type":"ContainerDied","Data":"7a90617b3b0d6771c977d6c48b45ea4ff4fc981c02f5e742dc62296448b86f5a"} Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.194254 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:12:04 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:12:04 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:12:04 crc kubenswrapper[4983]: healthz check failed Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.194565 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.211205 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563210-lptd5" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.295993 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kplv\" (UniqueName: \"kubernetes.io/projected/efeae5ef-dba6-4031-83be-8f493fe853b9-kube-api-access-5kplv\") pod \"efeae5ef-dba6-4031-83be-8f493fe853b9\" (UID: \"efeae5ef-dba6-4031-83be-8f493fe853b9\") " Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.310092 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efeae5ef-dba6-4031-83be-8f493fe853b9-kube-api-access-5kplv" (OuterVolumeSpecName: "kube-api-access-5kplv") pod "efeae5ef-dba6-4031-83be-8f493fe853b9" (UID: "efeae5ef-dba6-4031-83be-8f493fe853b9"). InnerVolumeSpecName "kube-api-access-5kplv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.342129 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.353427 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.397278 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef-kubelet-dir\") pod \"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef\" (UID: \"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef\") " Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.397419 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2bab65a6-74ae-4323-a9a8-bc3da3097d6f-kubelet-dir\") pod \"2bab65a6-74ae-4323-a9a8-bc3da3097d6f\" (UID: \"2bab65a6-74ae-4323-a9a8-bc3da3097d6f\") " Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.397477 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef" (UID: "ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.397498 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef-kube-api-access\") pod \"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef\" (UID: \"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef\") " Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.397612 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2bab65a6-74ae-4323-a9a8-bc3da3097d6f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2bab65a6-74ae-4323-a9a8-bc3da3097d6f" (UID: "2bab65a6-74ae-4323-a9a8-bc3da3097d6f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.397652 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2bab65a6-74ae-4323-a9a8-bc3da3097d6f-kube-api-access\") pod \"2bab65a6-74ae-4323-a9a8-bc3da3097d6f\" (UID: \"2bab65a6-74ae-4323-a9a8-bc3da3097d6f\") " Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.398053 4983 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2bab65a6-74ae-4323-a9a8-bc3da3097d6f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.398073 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kplv\" (UniqueName: \"kubernetes.io/projected/efeae5ef-dba6-4031-83be-8f493fe853b9-kube-api-access-5kplv\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.398085 4983 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.400708 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef" (UID: "ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.400849 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bab65a6-74ae-4323-a9a8-bc3da3097d6f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2bab65a6-74ae-4323-a9a8-bc3da3097d6f" (UID: "2bab65a6-74ae-4323-a9a8-bc3da3097d6f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.499386 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.499416 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2bab65a6-74ae-4323-a9a8-bc3da3097d6f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.702230 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.704254 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.716431 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe22dcea-cd03-43e3-80da-61a5f1e063ad-metrics-certs\") pod \"network-metrics-daemon-vjcg5\" (UID: \"fe22dcea-cd03-43e3-80da-61a5f1e063ad\") " pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.880503 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.890290 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vjcg5" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.938244 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.938234 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2bab65a6-74ae-4323-a9a8-bc3da3097d6f","Type":"ContainerDied","Data":"1e0ad35f2ad6c446e25d7a5af4a6c7038030b2c1b3f1073d325f758d94382e7b"} Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.938382 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e0ad35f2ad6c446e25d7a5af4a6c7038030b2c1b3f1073d325f758d94382e7b" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.948558 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.948550 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef","Type":"ContainerDied","Data":"2c6a6c9cb65162fc4c7289ba726f1fd801cd5808cec47702dff17e8d0676dc47"} Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.948688 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c6a6c9cb65162fc4c7289ba726f1fd801cd5808cec47702dff17e8d0676dc47" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.963463 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563210-lptd5" event={"ID":"efeae5ef-dba6-4031-83be-8f493fe853b9","Type":"ContainerDied","Data":"48c5b64768b2d14183c7a618f61e947bc1f88fa07119bdef7b88bd1fc4adb3ed"} Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.963525 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48c5b64768b2d14183c7a618f61e947bc1f88fa07119bdef7b88bd1fc4adb3ed" Mar 18 00:12:04 crc kubenswrapper[4983]: I0318 00:12:04.963614 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563210-lptd5" Mar 18 00:12:05 crc kubenswrapper[4983]: I0318 00:12:05.195351 4983 patch_prober.go:28] interesting pod/router-default-5444994796-ncss9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 00:12:05 crc kubenswrapper[4983]: [-]has-synced failed: reason withheld Mar 18 00:12:05 crc kubenswrapper[4983]: [+]process-running ok Mar 18 00:12:05 crc kubenswrapper[4983]: healthz check failed Mar 18 00:12:05 crc kubenswrapper[4983]: I0318 00:12:05.195744 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ncss9" podUID="dafd0e48-880c-44fc-b8c5-8056a37dac27" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 00:12:05 crc kubenswrapper[4983]: I0318 00:12:05.285141 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563212-g7qg2" Mar 18 00:12:05 crc kubenswrapper[4983]: I0318 00:12:05.309901 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btznr\" (UniqueName: \"kubernetes.io/projected/55e41d80-a523-4a82-8d42-c01daa50d832-kube-api-access-btznr\") pod \"55e41d80-a523-4a82-8d42-c01daa50d832\" (UID: \"55e41d80-a523-4a82-8d42-c01daa50d832\") " Mar 18 00:12:05 crc kubenswrapper[4983]: I0318 00:12:05.316132 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55e41d80-a523-4a82-8d42-c01daa50d832-kube-api-access-btznr" (OuterVolumeSpecName: "kube-api-access-btznr") pod "55e41d80-a523-4a82-8d42-c01daa50d832" (UID: "55e41d80-a523-4a82-8d42-c01daa50d832"). InnerVolumeSpecName "kube-api-access-btznr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:05 crc kubenswrapper[4983]: I0318 00:12:05.411351 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btznr\" (UniqueName: \"kubernetes.io/projected/55e41d80-a523-4a82-8d42-c01daa50d832-kube-api-access-btznr\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:05 crc kubenswrapper[4983]: I0318 00:12:05.473053 4983 patch_prober.go:28] interesting pod/console-f9d7485db-65l25 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 18 00:12:05 crc kubenswrapper[4983]: I0318 00:12:05.473108 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-65l25" podUID="6375157a-012b-441e-b97b-39d6830cef68" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 18 00:12:05 crc kubenswrapper[4983]: I0318 00:12:05.587241 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vjcg5"] Mar 18 00:12:06 crc kubenswrapper[4983]: I0318 00:12:05.997425 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563212-g7qg2" event={"ID":"55e41d80-a523-4a82-8d42-c01daa50d832","Type":"ContainerDied","Data":"c741f78e43e4310dede5dd1f652f3cd709e87461bc1bc84ce80accdb862f9529"} Mar 18 00:12:06 crc kubenswrapper[4983]: I0318 00:12:05.997680 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c741f78e43e4310dede5dd1f652f3cd709e87461bc1bc84ce80accdb862f9529" Mar 18 00:12:06 crc kubenswrapper[4983]: I0318 00:12:05.997752 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563212-g7qg2" Mar 18 00:12:06 crc kubenswrapper[4983]: I0318 00:12:06.145333 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-87bh5" Mar 18 00:12:06 crc kubenswrapper[4983]: I0318 00:12:06.194513 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:12:06 crc kubenswrapper[4983]: I0318 00:12:06.196375 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-ncss9" Mar 18 00:12:10 crc kubenswrapper[4983]: I0318 00:12:10.840165 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-57d94b7d77-crjwh"] Mar 18 00:12:10 crc kubenswrapper[4983]: I0318 00:12:10.842609 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" podUID="c27c1f03-b245-4beb-9de4-f17cb0844f7a" containerName="controller-manager" containerID="cri-o://d72721315f30ae7d13f4a96c5cc25b8b3599a375f273fcdcedce84e09b1c0249" gracePeriod=30 Mar 18 00:12:10 crc kubenswrapper[4983]: I0318 00:12:10.864812 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb"] Mar 18 00:12:10 crc kubenswrapper[4983]: I0318 00:12:10.865072 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" podUID="0c616674-2247-42dd-9079-0ff8906d30da" containerName="route-controller-manager" containerID="cri-o://33795ce60bdb6d48e29737e87fc9580caacc10c25d0ff4334bd8828fdce2856e" gracePeriod=30 Mar 18 00:12:12 crc kubenswrapper[4983]: I0318 00:12:12.042261 4983 generic.go:334] "Generic (PLEG): container finished" podID="c27c1f03-b245-4beb-9de4-f17cb0844f7a" containerID="d72721315f30ae7d13f4a96c5cc25b8b3599a375f273fcdcedce84e09b1c0249" exitCode=0 Mar 18 00:12:12 crc kubenswrapper[4983]: I0318 00:12:12.042401 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" event={"ID":"c27c1f03-b245-4beb-9de4-f17cb0844f7a","Type":"ContainerDied","Data":"d72721315f30ae7d13f4a96c5cc25b8b3599a375f273fcdcedce84e09b1c0249"} Mar 18 00:12:12 crc kubenswrapper[4983]: I0318 00:12:12.043962 4983 generic.go:334] "Generic (PLEG): container finished" podID="0c616674-2247-42dd-9079-0ff8906d30da" containerID="33795ce60bdb6d48e29737e87fc9580caacc10c25d0ff4334bd8828fdce2856e" exitCode=0 Mar 18 00:12:12 crc kubenswrapper[4983]: I0318 00:12:12.044006 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" event={"ID":"0c616674-2247-42dd-9079-0ff8906d30da","Type":"ContainerDied","Data":"33795ce60bdb6d48e29737e87fc9580caacc10c25d0ff4334bd8828fdce2856e"} Mar 18 00:12:13 crc kubenswrapper[4983]: I0318 00:12:13.259831 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:12:13 crc kubenswrapper[4983]: I0318 00:12:13.259903 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:12:13 crc kubenswrapper[4983]: I0318 00:12:13.276257 4983 patch_prober.go:28] interesting pod/controller-manager-57d94b7d77-crjwh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.47:8443/healthz\": dial tcp 10.217.0.47:8443: connect: connection refused" start-of-body= Mar 18 00:12:13 crc kubenswrapper[4983]: I0318 00:12:13.276313 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" podUID="c27c1f03-b245-4beb-9de4-f17cb0844f7a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.47:8443/healthz\": dial tcp 10.217.0.47:8443: connect: connection refused" Mar 18 00:12:15 crc kubenswrapper[4983]: I0318 00:12:15.478328 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:12:15 crc kubenswrapper[4983]: I0318 00:12:15.484217 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-65l25" Mar 18 00:12:15 crc kubenswrapper[4983]: I0318 00:12:15.776035 4983 patch_prober.go:28] interesting pod/route-controller-manager-58ff7bcd8-w5scb container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" start-of-body= Mar 18 00:12:15 crc kubenswrapper[4983]: I0318 00:12:15.776098 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" podUID="0c616674-2247-42dd-9079-0ff8906d30da" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" Mar 18 00:12:16 crc kubenswrapper[4983]: I0318 00:12:16.002257 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 00:12:17 crc kubenswrapper[4983]: I0318 00:12:17.080374 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" event={"ID":"fe22dcea-cd03-43e3-80da-61a5f1e063ad","Type":"ContainerStarted","Data":"45d851a034a05128482fd8dc17940d78a9ee76d4f6b49ffb45d8aafb5c930616"} Mar 18 00:12:17 crc kubenswrapper[4983]: I0318 00:12:17.559476 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:12:20 crc kubenswrapper[4983]: I0318 00:12:20.096555 4983 generic.go:334] "Generic (PLEG): container finished" podID="c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8" containerID="b9dfc5a4c8fa5a62b5be3aa151ae19a10c0341e65307da84c928ebcfd7457397" exitCode=0 Mar 18 00:12:20 crc kubenswrapper[4983]: I0318 00:12:20.096860 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29563200-6l8ds" event={"ID":"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8","Type":"ContainerDied","Data":"b9dfc5a4c8fa5a62b5be3aa151ae19a10c0341e65307da84c928ebcfd7457397"} Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.878244 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.909984 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw"] Mar 18 00:12:21 crc kubenswrapper[4983]: E0318 00:12:21.910196 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bab65a6-74ae-4323-a9a8-bc3da3097d6f" containerName="pruner" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.910213 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bab65a6-74ae-4323-a9a8-bc3da3097d6f" containerName="pruner" Mar 18 00:12:21 crc kubenswrapper[4983]: E0318 00:12:21.910230 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55e41d80-a523-4a82-8d42-c01daa50d832" containerName="oc" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.910237 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="55e41d80-a523-4a82-8d42-c01daa50d832" containerName="oc" Mar 18 00:12:21 crc kubenswrapper[4983]: E0318 00:12:21.910249 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c616674-2247-42dd-9079-0ff8906d30da" containerName="route-controller-manager" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.910255 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c616674-2247-42dd-9079-0ff8906d30da" containerName="route-controller-manager" Mar 18 00:12:21 crc kubenswrapper[4983]: E0318 00:12:21.910263 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef" containerName="pruner" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.910269 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef" containerName="pruner" Mar 18 00:12:21 crc kubenswrapper[4983]: E0318 00:12:21.910278 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efeae5ef-dba6-4031-83be-8f493fe853b9" containerName="oc" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.910284 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="efeae5ef-dba6-4031-83be-8f493fe853b9" containerName="oc" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.910380 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="efeae5ef-dba6-4031-83be-8f493fe853b9" containerName="oc" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.910390 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff580ae7-a651-4c7f-ad16-8a0a3ccfbeef" containerName="pruner" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.910404 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="55e41d80-a523-4a82-8d42-c01daa50d832" containerName="oc" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.910412 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c616674-2247-42dd-9079-0ff8906d30da" containerName="route-controller-manager" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.910419 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bab65a6-74ae-4323-a9a8-bc3da3097d6f" containerName="pruner" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.910756 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.920532 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw"] Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.963589 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c616674-2247-42dd-9079-0ff8906d30da-serving-cert\") pod \"0c616674-2247-42dd-9079-0ff8906d30da\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.963679 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rghv4\" (UniqueName: \"kubernetes.io/projected/0c616674-2247-42dd-9079-0ff8906d30da-kube-api-access-rghv4\") pod \"0c616674-2247-42dd-9079-0ff8906d30da\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.963734 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c616674-2247-42dd-9079-0ff8906d30da-client-ca\") pod \"0c616674-2247-42dd-9079-0ff8906d30da\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.963775 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c616674-2247-42dd-9079-0ff8906d30da-config\") pod \"0c616674-2247-42dd-9079-0ff8906d30da\" (UID: \"0c616674-2247-42dd-9079-0ff8906d30da\") " Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.964672 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c616674-2247-42dd-9079-0ff8906d30da-config" (OuterVolumeSpecName: "config") pod "0c616674-2247-42dd-9079-0ff8906d30da" (UID: "0c616674-2247-42dd-9079-0ff8906d30da"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.965034 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c616674-2247-42dd-9079-0ff8906d30da-client-ca" (OuterVolumeSpecName: "client-ca") pod "0c616674-2247-42dd-9079-0ff8906d30da" (UID: "0c616674-2247-42dd-9079-0ff8906d30da"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.977070 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c616674-2247-42dd-9079-0ff8906d30da-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0c616674-2247-42dd-9079-0ff8906d30da" (UID: "0c616674-2247-42dd-9079-0ff8906d30da"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:12:21 crc kubenswrapper[4983]: I0318 00:12:21.983590 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c616674-2247-42dd-9079-0ff8906d30da-kube-api-access-rghv4" (OuterVolumeSpecName: "kube-api-access-rghv4") pod "0c616674-2247-42dd-9079-0ff8906d30da" (UID: "0c616674-2247-42dd-9079-0ff8906d30da"). InnerVolumeSpecName "kube-api-access-rghv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.065764 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-client-ca\") pod \"route-controller-manager-cc887d947-76ppw\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.065858 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-config\") pod \"route-controller-manager-cc887d947-76ppw\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.065889 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpbzh\" (UniqueName: \"kubernetes.io/projected/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-kube-api-access-xpbzh\") pod \"route-controller-manager-cc887d947-76ppw\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.065940 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-serving-cert\") pod \"route-controller-manager-cc887d947-76ppw\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.065989 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c616674-2247-42dd-9079-0ff8906d30da-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.066001 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rghv4\" (UniqueName: \"kubernetes.io/projected/0c616674-2247-42dd-9079-0ff8906d30da-kube-api-access-rghv4\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.066009 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c616674-2247-42dd-9079-0ff8906d30da-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.066017 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c616674-2247-42dd-9079-0ff8906d30da-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.107877 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" event={"ID":"0c616674-2247-42dd-9079-0ff8906d30da","Type":"ContainerDied","Data":"f387cfa53894e2bf23dcc323f4ca61dc029f8fa193ecedcf1ed6123d414bc8a9"} Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.107949 4983 scope.go:117] "RemoveContainer" containerID="33795ce60bdb6d48e29737e87fc9580caacc10c25d0ff4334bd8828fdce2856e" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.107968 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.134848 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb"] Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.137896 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58ff7bcd8-w5scb"] Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.171444 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-config\") pod \"route-controller-manager-cc887d947-76ppw\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.171511 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpbzh\" (UniqueName: \"kubernetes.io/projected/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-kube-api-access-xpbzh\") pod \"route-controller-manager-cc887d947-76ppw\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.171590 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-serving-cert\") pod \"route-controller-manager-cc887d947-76ppw\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.171656 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-client-ca\") pod \"route-controller-manager-cc887d947-76ppw\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.173005 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-client-ca\") pod \"route-controller-manager-cc887d947-76ppw\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.175671 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-config\") pod \"route-controller-manager-cc887d947-76ppw\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.179185 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-serving-cert\") pod \"route-controller-manager-cc887d947-76ppw\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.187124 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpbzh\" (UniqueName: \"kubernetes.io/projected/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-kube-api-access-xpbzh\") pod \"route-controller-manager-cc887d947-76ppw\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.237671 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:22 crc kubenswrapper[4983]: I0318 00:12:22.850408 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c616674-2247-42dd-9079-0ff8906d30da" path="/var/lib/kubelet/pods/0c616674-2247-42dd-9079-0ff8906d30da/volumes" Mar 18 00:12:23 crc kubenswrapper[4983]: I0318 00:12:23.277051 4983 patch_prober.go:28] interesting pod/controller-manager-57d94b7d77-crjwh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.47:8443/healthz\": dial tcp 10.217.0.47:8443: connect: connection refused" start-of-body= Mar 18 00:12:23 crc kubenswrapper[4983]: I0318 00:12:23.277133 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" podUID="c27c1f03-b245-4beb-9de4-f17cb0844f7a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.47:8443/healthz\": dial tcp 10.217.0.47:8443: connect: connection refused" Mar 18 00:12:26 crc kubenswrapper[4983]: I0318 00:12:26.509523 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-96hcz" Mar 18 00:12:27 crc kubenswrapper[4983]: I0318 00:12:27.731853 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6t6lv"] Mar 18 00:12:27 crc kubenswrapper[4983]: I0318 00:12:27.902019 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29563200-6l8ds" Mar 18 00:12:27 crc kubenswrapper[4983]: I0318 00:12:27.983222 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.045294 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2ms6\" (UniqueName: \"kubernetes.io/projected/c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8-kube-api-access-t2ms6\") pod \"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8\" (UID: \"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8\") " Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.045342 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8-serviceca\") pod \"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8\" (UID: \"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8\") " Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.046964 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8-serviceca" (OuterVolumeSpecName: "serviceca") pod "c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8" (UID: "c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.059371 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8-kube-api-access-t2ms6" (OuterVolumeSpecName: "kube-api-access-t2ms6") pod "c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8" (UID: "c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8"). InnerVolumeSpecName "kube-api-access-t2ms6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.145425 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29563200-6l8ds" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.145437 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29563200-6l8ds" event={"ID":"c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8","Type":"ContainerDied","Data":"3fce4956773ff2e831fd8c56997a98538f386e09f45740556fa4487b7ea44a83"} Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.145787 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fce4956773ff2e831fd8c56997a98538f386e09f45740556fa4487b7ea44a83" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.146109 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-config\") pod \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.146144 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c27c1f03-b245-4beb-9de4-f17cb0844f7a-serving-cert\") pod \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.146212 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9q8l\" (UniqueName: \"kubernetes.io/projected/c27c1f03-b245-4beb-9de4-f17cb0844f7a-kube-api-access-v9q8l\") pod \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.146240 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-client-ca\") pod \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.146308 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-proxy-ca-bundles\") pod \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\" (UID: \"c27c1f03-b245-4beb-9de4-f17cb0844f7a\") " Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.146502 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2ms6\" (UniqueName: \"kubernetes.io/projected/c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8-kube-api-access-t2ms6\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.146514 4983 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8-serviceca\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.147206 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c27c1f03-b245-4beb-9de4-f17cb0844f7a" (UID: "c27c1f03-b245-4beb-9de4-f17cb0844f7a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.147296 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-config" (OuterVolumeSpecName: "config") pod "c27c1f03-b245-4beb-9de4-f17cb0844f7a" (UID: "c27c1f03-b245-4beb-9de4-f17cb0844f7a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.147276 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-client-ca" (OuterVolumeSpecName: "client-ca") pod "c27c1f03-b245-4beb-9de4-f17cb0844f7a" (UID: "c27c1f03-b245-4beb-9de4-f17cb0844f7a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.150455 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c27c1f03-b245-4beb-9de4-f17cb0844f7a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c27c1f03-b245-4beb-9de4-f17cb0844f7a" (UID: "c27c1f03-b245-4beb-9de4-f17cb0844f7a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.151757 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.152010 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57d94b7d77-crjwh" event={"ID":"c27c1f03-b245-4beb-9de4-f17cb0844f7a","Type":"ContainerDied","Data":"6dbbad3fd96b6fb449021e881645bee81a8cbca2256b558a8b2fdd2f9ad0778a"} Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.152055 4983 scope.go:117] "RemoveContainer" containerID="d72721315f30ae7d13f4a96c5cc25b8b3599a375f273fcdcedce84e09b1c0249" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.152287 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c27c1f03-b245-4beb-9de4-f17cb0844f7a-kube-api-access-v9q8l" (OuterVolumeSpecName: "kube-api-access-v9q8l") pod "c27c1f03-b245-4beb-9de4-f17cb0844f7a" (UID: "c27c1f03-b245-4beb-9de4-f17cb0844f7a"). InnerVolumeSpecName "kube-api-access-v9q8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.234608 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw"] Mar 18 00:12:28 crc kubenswrapper[4983]: W0318 00:12:28.236386 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0cf6e6a_58ce_45fd_a802_1ade4a07bdac.slice/crio-988fba98614713cefecab16bb1c4ff3a0ce6e666bf9431aa9260bf6dcb42786e WatchSource:0}: Error finding container 988fba98614713cefecab16bb1c4ff3a0ce6e666bf9431aa9260bf6dcb42786e: Status 404 returned error can't find the container with id 988fba98614713cefecab16bb1c4ff3a0ce6e666bf9431aa9260bf6dcb42786e Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.247523 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.247554 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c27c1f03-b245-4beb-9de4-f17cb0844f7a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.247568 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9q8l\" (UniqueName: \"kubernetes.io/projected/c27c1f03-b245-4beb-9de4-f17cb0844f7a-kube-api-access-v9q8l\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.247579 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.247587 4983 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c27c1f03-b245-4beb-9de4-f17cb0844f7a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.517596 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-57d94b7d77-crjwh"] Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.517855 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-57d94b7d77-crjwh"] Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.673497 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 18 00:12:28 crc kubenswrapper[4983]: E0318 00:12:28.673701 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c27c1f03-b245-4beb-9de4-f17cb0844f7a" containerName="controller-manager" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.673712 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="c27c1f03-b245-4beb-9de4-f17cb0844f7a" containerName="controller-manager" Mar 18 00:12:28 crc kubenswrapper[4983]: E0318 00:12:28.673724 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8" containerName="image-pruner" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.673729 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8" containerName="image-pruner" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.673824 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="c27c1f03-b245-4beb-9de4-f17cb0844f7a" containerName="controller-manager" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.673834 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3c4014a-24f4-4e23-8d99-0a6ffabeb4a8" containerName="image-pruner" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.674172 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.679837 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.680130 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.684867 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.758631 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/386588e9-85fe-403d-a200-da43cf4cd6da-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"386588e9-85fe-403d-a200-da43cf4cd6da\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.758695 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/386588e9-85fe-403d-a200-da43cf4cd6da-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"386588e9-85fe-403d-a200-da43cf4cd6da\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.856152 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c27c1f03-b245-4beb-9de4-f17cb0844f7a" path="/var/lib/kubelet/pods/c27c1f03-b245-4beb-9de4-f17cb0844f7a/volumes" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.859427 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/386588e9-85fe-403d-a200-da43cf4cd6da-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"386588e9-85fe-403d-a200-da43cf4cd6da\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.859487 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/386588e9-85fe-403d-a200-da43cf4cd6da-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"386588e9-85fe-403d-a200-da43cf4cd6da\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.859591 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/386588e9-85fe-403d-a200-da43cf4cd6da-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"386588e9-85fe-403d-a200-da43cf4cd6da\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 00:12:28 crc kubenswrapper[4983]: I0318 00:12:28.880188 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/386588e9-85fe-403d-a200-da43cf4cd6da-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"386588e9-85fe-403d-a200-da43cf4cd6da\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.023262 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.166561 4983 generic.go:334] "Generic (PLEG): container finished" podID="089eb8e9-84d9-4320-97ec-9d868b858d3a" containerID="a8c73084999085c4d55387181fe5c17114aa76abc0ff6b47f2d763fbda787e55" exitCode=0 Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.167054 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fs978" event={"ID":"089eb8e9-84d9-4320-97ec-9d868b858d3a","Type":"ContainerDied","Data":"a8c73084999085c4d55387181fe5c17114aa76abc0ff6b47f2d763fbda787e55"} Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.171851 4983 generic.go:334] "Generic (PLEG): container finished" podID="a81a1b18-1976-45f7-97f5-4dcd7f745af3" containerID="2fa0c0dbe2e1b40055a0e3bac3a783bbaace18295265c106a565a4ea2e8a7856" exitCode=0 Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.171892 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnc92" event={"ID":"a81a1b18-1976-45f7-97f5-4dcd7f745af3","Type":"ContainerDied","Data":"2fa0c0dbe2e1b40055a0e3bac3a783bbaace18295265c106a565a4ea2e8a7856"} Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.175548 4983 generic.go:334] "Generic (PLEG): container finished" podID="fe54f632-0209-4996-979c-4392b8f7d037" containerID="8c1f9d5a50870abb84e98864f555ff379749fb5600491b30101a9862a989efd9" exitCode=0 Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.175586 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fpkl" event={"ID":"fe54f632-0209-4996-979c-4392b8f7d037","Type":"ContainerDied","Data":"8c1f9d5a50870abb84e98864f555ff379749fb5600491b30101a9862a989efd9"} Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.192753 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" event={"ID":"fe22dcea-cd03-43e3-80da-61a5f1e063ad","Type":"ContainerStarted","Data":"3933f3009313b3b0865c99603cdbc6304ed8e1e439d0241cdb1f25ee74ca8199"} Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.192798 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vjcg5" event={"ID":"fe22dcea-cd03-43e3-80da-61a5f1e063ad","Type":"ContainerStarted","Data":"a02fde2bfeaf72e00d9aa47a7103c04b840fdf6b17410e2f5abfe1d4f86c7504"} Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.197453 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78tfj" event={"ID":"75077192-24a6-4efb-8484-bea2db4a4812","Type":"ContainerStarted","Data":"f28a8dc1c6c6da99321c8222daf8d5a4fe3e624a5670cc39e3073c6d7e59035a"} Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.199900 4983 generic.go:334] "Generic (PLEG): container finished" podID="5099d24e-b4a3-4990-bc90-980e0188ef3e" containerID="78e8203f38fb30d66ba733d9b011647d48d43c26289504b858d26c163f1db6ac" exitCode=0 Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.200726 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lqjwd" event={"ID":"5099d24e-b4a3-4990-bc90-980e0188ef3e","Type":"ContainerDied","Data":"78e8203f38fb30d66ba733d9b011647d48d43c26289504b858d26c163f1db6ac"} Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.210100 4983 generic.go:334] "Generic (PLEG): container finished" podID="41c73828-774d-4440-a31d-c7cca08b063e" containerID="104f78936c8abfe492c6fa4a7e951f6dde1c4f8d8d79f7f62f1653409d12bf31" exitCode=0 Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.210158 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxm7g" event={"ID":"41c73828-774d-4440-a31d-c7cca08b063e","Type":"ContainerDied","Data":"104f78936c8abfe492c6fa4a7e951f6dde1c4f8d8d79f7f62f1653409d12bf31"} Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.217339 4983 generic.go:334] "Generic (PLEG): container finished" podID="b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" containerID="684451e4702221db75489b9ec2b8673dc727d8bd912649ca7c5d51787777f1dd" exitCode=0 Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.217375 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdsf7" event={"ID":"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419","Type":"ContainerDied","Data":"684451e4702221db75489b9ec2b8673dc727d8bd912649ca7c5d51787777f1dd"} Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.223444 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5mgj" event={"ID":"5e139d3f-fd2e-4e39-96b4-eafd340c10a3","Type":"ContainerStarted","Data":"d9dfb1511dc0a58eacc6af7b10f950487159392752cdfd91ae4dffb5530c4b59"} Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.230456 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" event={"ID":"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac","Type":"ContainerStarted","Data":"5bc85467d3a0e9bc4f0ba7715a8856037f8b9f6a65f7425a3cca168c7e5dd228"} Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.230486 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" event={"ID":"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac","Type":"ContainerStarted","Data":"988fba98614713cefecab16bb1c4ff3a0ce6e666bf9431aa9260bf6dcb42786e"} Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.230967 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.237545 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.245139 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:29 crc kubenswrapper[4983]: I0318 00:12:29.332760 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" podStartSLOduration=19.332745552 podStartE2EDuration="19.332745552s" podCreationTimestamp="2026-03-18 00:12:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:12:29.329457763 +0000 UTC m=+267.226832383" watchObservedRunningTime="2026-03-18 00:12:29.332745552 +0000 UTC m=+267.230120162" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.236756 4983 generic.go:334] "Generic (PLEG): container finished" podID="75077192-24a6-4efb-8484-bea2db4a4812" containerID="f28a8dc1c6c6da99321c8222daf8d5a4fe3e624a5670cc39e3073c6d7e59035a" exitCode=0 Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.236851 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78tfj" event={"ID":"75077192-24a6-4efb-8484-bea2db4a4812","Type":"ContainerDied","Data":"f28a8dc1c6c6da99321c8222daf8d5a4fe3e624a5670cc39e3073c6d7e59035a"} Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.239721 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"386588e9-85fe-403d-a200-da43cf4cd6da","Type":"ContainerStarted","Data":"589bfd6930bdc41c81cd2e1929f70f24dbc0d2e1c51620613d91af6529b4356a"} Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.239761 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"386588e9-85fe-403d-a200-da43cf4cd6da","Type":"ContainerStarted","Data":"dda150b77426baf70247e2a483e9c5434ed86ac67a4ab1056bc2782cabb21d52"} Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.242481 4983 generic.go:334] "Generic (PLEG): container finished" podID="5e139d3f-fd2e-4e39-96b4-eafd340c10a3" containerID="d9dfb1511dc0a58eacc6af7b10f950487159392752cdfd91ae4dffb5530c4b59" exitCode=0 Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.243469 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5mgj" event={"ID":"5e139d3f-fd2e-4e39-96b4-eafd340c10a3","Type":"ContainerDied","Data":"d9dfb1511dc0a58eacc6af7b10f950487159392752cdfd91ae4dffb5530c4b59"} Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.284602 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.284580063 podStartE2EDuration="2.284580063s" podCreationTimestamp="2026-03-18 00:12:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:12:30.284220462 +0000 UTC m=+268.181595072" watchObservedRunningTime="2026-03-18 00:12:30.284580063 +0000 UTC m=+268.181954693" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.300638 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-vjcg5" podStartSLOduration=206.300621529 podStartE2EDuration="3m26.300621529s" podCreationTimestamp="2026-03-18 00:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:12:30.295851514 +0000 UTC m=+268.193226124" watchObservedRunningTime="2026-03-18 00:12:30.300621529 +0000 UTC m=+268.197996139" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.886059 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5b768657c7-9dlrt"] Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.887209 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.888636 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.890508 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.890723 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.890794 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.891053 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.891311 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.901393 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b768657c7-9dlrt"] Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.914148 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.947863 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw"] Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.987460 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2b8r\" (UniqueName: \"kubernetes.io/projected/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-kube-api-access-w2b8r\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.987537 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-proxy-ca-bundles\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.987603 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-client-ca\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.987630 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-serving-cert\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:30 crc kubenswrapper[4983]: I0318 00:12:30.987668 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-config\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.088686 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-serving-cert\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.088732 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-config\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.088787 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2b8r\" (UniqueName: \"kubernetes.io/projected/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-kube-api-access-w2b8r\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.088822 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-proxy-ca-bundles\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.088849 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-client-ca\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.089862 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-client-ca\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.090466 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-proxy-ca-bundles\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.090612 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-config\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.098183 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-serving-cert\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.103714 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2b8r\" (UniqueName: \"kubernetes.io/projected/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-kube-api-access-w2b8r\") pod \"controller-manager-5b768657c7-9dlrt\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.239052 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.249045 4983 generic.go:334] "Generic (PLEG): container finished" podID="386588e9-85fe-403d-a200-da43cf4cd6da" containerID="589bfd6930bdc41c81cd2e1929f70f24dbc0d2e1c51620613d91af6529b4356a" exitCode=0 Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.249212 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"386588e9-85fe-403d-a200-da43cf4cd6da","Type":"ContainerDied","Data":"589bfd6930bdc41c81cd2e1929f70f24dbc0d2e1c51620613d91af6529b4356a"} Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.250980 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fpkl" event={"ID":"fe54f632-0209-4996-979c-4392b8f7d037","Type":"ContainerStarted","Data":"2d6786e68a747e5bc51e202186f8f257e5b430d111b14ffabecb02fccc350f76"} Mar 18 00:12:31 crc kubenswrapper[4983]: I0318 00:12:31.881457 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b768657c7-9dlrt"] Mar 18 00:12:32 crc kubenswrapper[4983]: I0318 00:12:32.257775 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" event={"ID":"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f","Type":"ContainerStarted","Data":"f5ceab20ccd608a5ac263b84b4f311277d5bd49d71b0a0536143184d96cb10e9"} Mar 18 00:12:32 crc kubenswrapper[4983]: I0318 00:12:32.257848 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" podUID="e0cf6e6a-58ce-45fd-a802-1ade4a07bdac" containerName="route-controller-manager" containerID="cri-o://5bc85467d3a0e9bc4f0ba7715a8856037f8b9f6a65f7425a3cca168c7e5dd228" gracePeriod=30 Mar 18 00:12:32 crc kubenswrapper[4983]: I0318 00:12:32.278705 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8fpkl" podStartSLOduration=8.241111701 podStartE2EDuration="37.278683403s" podCreationTimestamp="2026-03-18 00:11:55 +0000 UTC" firstStartedPulling="2026-03-18 00:12:01.799616555 +0000 UTC m=+239.696991165" lastFinishedPulling="2026-03-18 00:12:30.837188257 +0000 UTC m=+268.734562867" observedRunningTime="2026-03-18 00:12:32.27691556 +0000 UTC m=+270.174290180" watchObservedRunningTime="2026-03-18 00:12:32.278683403 +0000 UTC m=+270.176058023" Mar 18 00:12:32 crc kubenswrapper[4983]: I0318 00:12:32.554802 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 00:12:32 crc kubenswrapper[4983]: I0318 00:12:32.707625 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/386588e9-85fe-403d-a200-da43cf4cd6da-kubelet-dir\") pod \"386588e9-85fe-403d-a200-da43cf4cd6da\" (UID: \"386588e9-85fe-403d-a200-da43cf4cd6da\") " Mar 18 00:12:32 crc kubenswrapper[4983]: I0318 00:12:32.707733 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/386588e9-85fe-403d-a200-da43cf4cd6da-kube-api-access\") pod \"386588e9-85fe-403d-a200-da43cf4cd6da\" (UID: \"386588e9-85fe-403d-a200-da43cf4cd6da\") " Mar 18 00:12:32 crc kubenswrapper[4983]: I0318 00:12:32.707771 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/386588e9-85fe-403d-a200-da43cf4cd6da-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "386588e9-85fe-403d-a200-da43cf4cd6da" (UID: "386588e9-85fe-403d-a200-da43cf4cd6da"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:12:32 crc kubenswrapper[4983]: I0318 00:12:32.708003 4983 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/386588e9-85fe-403d-a200-da43cf4cd6da-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:32 crc kubenswrapper[4983]: I0318 00:12:32.713113 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/386588e9-85fe-403d-a200-da43cf4cd6da-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "386588e9-85fe-403d-a200-da43cf4cd6da" (UID: "386588e9-85fe-403d-a200-da43cf4cd6da"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:32 crc kubenswrapper[4983]: I0318 00:12:32.809300 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/386588e9-85fe-403d-a200-da43cf4cd6da-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.266724 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"386588e9-85fe-403d-a200-da43cf4cd6da","Type":"ContainerDied","Data":"dda150b77426baf70247e2a483e9c5434ed86ac67a4ab1056bc2782cabb21d52"} Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.266760 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.266772 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dda150b77426baf70247e2a483e9c5434ed86ac67a4ab1056bc2782cabb21d52" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.268588 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" event={"ID":"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f","Type":"ContainerStarted","Data":"5f498d046341130439546e497a8bc3606487d4187f16d59dcdc8c2faf2cd4ad0"} Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.271521 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.276721 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.283055 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fs978" event={"ID":"089eb8e9-84d9-4320-97ec-9d868b858d3a","Type":"ContainerStarted","Data":"d2f46f717106369e8aad5052ffd0791f4b0713776fc894fee3b6827e42ea00a5"} Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.291616 4983 generic.go:334] "Generic (PLEG): container finished" podID="e0cf6e6a-58ce-45fd-a802-1ade4a07bdac" containerID="5bc85467d3a0e9bc4f0ba7715a8856037f8b9f6a65f7425a3cca168c7e5dd228" exitCode=0 Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.291861 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" event={"ID":"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac","Type":"ContainerDied","Data":"5bc85467d3a0e9bc4f0ba7715a8856037f8b9f6a65f7425a3cca168c7e5dd228"} Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.294189 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" podStartSLOduration=3.294150332 podStartE2EDuration="3.294150332s" podCreationTimestamp="2026-03-18 00:12:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:12:33.286816129 +0000 UTC m=+271.184190779" watchObservedRunningTime="2026-03-18 00:12:33.294150332 +0000 UTC m=+271.191524942" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.312879 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fs978" podStartSLOduration=6.222994952 podStartE2EDuration="35.312833538s" podCreationTimestamp="2026-03-18 00:11:58 +0000 UTC" firstStartedPulling="2026-03-18 00:12:02.828828769 +0000 UTC m=+240.726203379" lastFinishedPulling="2026-03-18 00:12:31.918667345 +0000 UTC m=+269.816041965" observedRunningTime="2026-03-18 00:12:33.307125235 +0000 UTC m=+271.204499855" watchObservedRunningTime="2026-03-18 00:12:33.312833538 +0000 UTC m=+271.210208148" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.409179 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.518201 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpbzh\" (UniqueName: \"kubernetes.io/projected/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-kube-api-access-xpbzh\") pod \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.518287 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-serving-cert\") pod \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.518325 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-client-ca\") pod \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.518349 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-config\") pod \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\" (UID: \"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac\") " Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.519124 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-client-ca" (OuterVolumeSpecName: "client-ca") pod "e0cf6e6a-58ce-45fd-a802-1ade4a07bdac" (UID: "e0cf6e6a-58ce-45fd-a802-1ade4a07bdac"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.519251 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-config" (OuterVolumeSpecName: "config") pod "e0cf6e6a-58ce-45fd-a802-1ade4a07bdac" (UID: "e0cf6e6a-58ce-45fd-a802-1ade4a07bdac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.536998 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e0cf6e6a-58ce-45fd-a802-1ade4a07bdac" (UID: "e0cf6e6a-58ce-45fd-a802-1ade4a07bdac"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.537221 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-kube-api-access-xpbzh" (OuterVolumeSpecName: "kube-api-access-xpbzh") pod "e0cf6e6a-58ce-45fd-a802-1ade4a07bdac" (UID: "e0cf6e6a-58ce-45fd-a802-1ade4a07bdac"). InnerVolumeSpecName "kube-api-access-xpbzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.619973 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpbzh\" (UniqueName: \"kubernetes.io/projected/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-kube-api-access-xpbzh\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.620007 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.620021 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:33 crc kubenswrapper[4983]: I0318 00:12:33.620031 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.301305 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxm7g" event={"ID":"41c73828-774d-4440-a31d-c7cca08b063e","Type":"ContainerStarted","Data":"bd9524e52f559d4990bcf31dc6a24772a6d37c53371a731f5bd2501a1892e07e"} Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.304159 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" event={"ID":"e0cf6e6a-58ce-45fd-a802-1ade4a07bdac","Type":"ContainerDied","Data":"988fba98614713cefecab16bb1c4ff3a0ce6e666bf9431aa9260bf6dcb42786e"} Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.304419 4983 scope.go:117] "RemoveContainer" containerID="5bc85467d3a0e9bc4f0ba7715a8856037f8b9f6a65f7425a3cca168c7e5dd228" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.304472 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.340830 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kxm7g" podStartSLOduration=7.601020684 podStartE2EDuration="38.340814036s" podCreationTimestamp="2026-03-18 00:11:56 +0000 UTC" firstStartedPulling="2026-03-18 00:12:02.835856129 +0000 UTC m=+240.733230729" lastFinishedPulling="2026-03-18 00:12:33.575649471 +0000 UTC m=+271.473024081" observedRunningTime="2026-03-18 00:12:34.323029987 +0000 UTC m=+272.220404597" watchObservedRunningTime="2026-03-18 00:12:34.340814036 +0000 UTC m=+272.238188646" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.343642 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw"] Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.346987 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cc887d947-76ppw"] Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.853635 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0cf6e6a-58ce-45fd-a802-1ade4a07bdac" path="/var/lib/kubelet/pods/e0cf6e6a-58ce-45fd-a802-1ade4a07bdac/volumes" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.864584 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv"] Mar 18 00:12:34 crc kubenswrapper[4983]: E0318 00:12:34.864834 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cf6e6a-58ce-45fd-a802-1ade4a07bdac" containerName="route-controller-manager" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.864853 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cf6e6a-58ce-45fd-a802-1ade4a07bdac" containerName="route-controller-manager" Mar 18 00:12:34 crc kubenswrapper[4983]: E0318 00:12:34.864867 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="386588e9-85fe-403d-a200-da43cf4cd6da" containerName="pruner" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.864875 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="386588e9-85fe-403d-a200-da43cf4cd6da" containerName="pruner" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.865018 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0cf6e6a-58ce-45fd-a802-1ade4a07bdac" containerName="route-controller-manager" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.865037 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="386588e9-85fe-403d-a200-da43cf4cd6da" containerName="pruner" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.865459 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.867334 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.868550 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.868731 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.868552 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.872409 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv"] Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.872519 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.872748 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.944581 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptkns\" (UniqueName: \"kubernetes.io/projected/a213d891-554c-4985-bc09-c03ab2e58198-kube-api-access-ptkns\") pod \"route-controller-manager-6b4dc48cdc-v45kv\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.944983 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a213d891-554c-4985-bc09-c03ab2e58198-config\") pod \"route-controller-manager-6b4dc48cdc-v45kv\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.945010 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a213d891-554c-4985-bc09-c03ab2e58198-serving-cert\") pod \"route-controller-manager-6b4dc48cdc-v45kv\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:34 crc kubenswrapper[4983]: I0318 00:12:34.945033 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a213d891-554c-4985-bc09-c03ab2e58198-client-ca\") pod \"route-controller-manager-6b4dc48cdc-v45kv\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.046120 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a213d891-554c-4985-bc09-c03ab2e58198-client-ca\") pod \"route-controller-manager-6b4dc48cdc-v45kv\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.046189 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptkns\" (UniqueName: \"kubernetes.io/projected/a213d891-554c-4985-bc09-c03ab2e58198-kube-api-access-ptkns\") pod \"route-controller-manager-6b4dc48cdc-v45kv\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.046254 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a213d891-554c-4985-bc09-c03ab2e58198-config\") pod \"route-controller-manager-6b4dc48cdc-v45kv\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.046271 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a213d891-554c-4985-bc09-c03ab2e58198-serving-cert\") pod \"route-controller-manager-6b4dc48cdc-v45kv\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.047538 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a213d891-554c-4985-bc09-c03ab2e58198-client-ca\") pod \"route-controller-manager-6b4dc48cdc-v45kv\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.047673 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a213d891-554c-4985-bc09-c03ab2e58198-config\") pod \"route-controller-manager-6b4dc48cdc-v45kv\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.056669 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a213d891-554c-4985-bc09-c03ab2e58198-serving-cert\") pod \"route-controller-manager-6b4dc48cdc-v45kv\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.069759 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptkns\" (UniqueName: \"kubernetes.io/projected/a213d891-554c-4985-bc09-c03ab2e58198-kube-api-access-ptkns\") pod \"route-controller-manager-6b4dc48cdc-v45kv\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.072329 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.072911 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.074593 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.075074 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.084289 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.201151 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.248155 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/551c788d-c0f2-4c95-b728-947250e387ef-var-lock\") pod \"installer-9-crc\" (UID: \"551c788d-c0f2-4c95-b728-947250e387ef\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.248292 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/551c788d-c0f2-4c95-b728-947250e387ef-kubelet-dir\") pod \"installer-9-crc\" (UID: \"551c788d-c0f2-4c95-b728-947250e387ef\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.248325 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/551c788d-c0f2-4c95-b728-947250e387ef-kube-api-access\") pod \"installer-9-crc\" (UID: \"551c788d-c0f2-4c95-b728-947250e387ef\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.312638 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnc92" event={"ID":"a81a1b18-1976-45f7-97f5-4dcd7f745af3","Type":"ContainerStarted","Data":"eda81797bff5f0f42dc4edec65d6455e9985956ab14c064f4e4970b15110ca7f"} Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.336370 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wnc92" podStartSLOduration=8.418083766 podStartE2EDuration="40.33635554s" podCreationTimestamp="2026-03-18 00:11:55 +0000 UTC" firstStartedPulling="2026-03-18 00:12:02.888679416 +0000 UTC m=+240.786054016" lastFinishedPulling="2026-03-18 00:12:34.80695118 +0000 UTC m=+272.704325790" observedRunningTime="2026-03-18 00:12:35.335874865 +0000 UTC m=+273.233249485" watchObservedRunningTime="2026-03-18 00:12:35.33635554 +0000 UTC m=+273.233730160" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.349604 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/551c788d-c0f2-4c95-b728-947250e387ef-var-lock\") pod \"installer-9-crc\" (UID: \"551c788d-c0f2-4c95-b728-947250e387ef\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.349682 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/551c788d-c0f2-4c95-b728-947250e387ef-kubelet-dir\") pod \"installer-9-crc\" (UID: \"551c788d-c0f2-4c95-b728-947250e387ef\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.349721 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/551c788d-c0f2-4c95-b728-947250e387ef-var-lock\") pod \"installer-9-crc\" (UID: \"551c788d-c0f2-4c95-b728-947250e387ef\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.349723 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/551c788d-c0f2-4c95-b728-947250e387ef-kube-api-access\") pod \"installer-9-crc\" (UID: \"551c788d-c0f2-4c95-b728-947250e387ef\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.349777 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/551c788d-c0f2-4c95-b728-947250e387ef-kubelet-dir\") pod \"installer-9-crc\" (UID: \"551c788d-c0f2-4c95-b728-947250e387ef\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.366431 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/551c788d-c0f2-4c95-b728-947250e387ef-kube-api-access\") pod \"installer-9-crc\" (UID: \"551c788d-c0f2-4c95-b728-947250e387ef\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.401879 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.923108 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:12:35 crc kubenswrapper[4983]: I0318 00:12:35.923790 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.100681 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.111922 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv"] Mar 18 00:12:36 crc kubenswrapper[4983]: W0318 00:12:36.118000 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda213d891_554c_4985_bc09_c03ab2e58198.slice/crio-5c6dcf4adaa0457cfad9919575d364a15ac1a8104cd8e631c71a743f8f24536f WatchSource:0}: Error finding container 5c6dcf4adaa0457cfad9919575d364a15ac1a8104cd8e631c71a743f8f24536f: Status 404 returned error can't find the container with id 5c6dcf4adaa0457cfad9919575d364a15ac1a8104cd8e631c71a743f8f24536f Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.159155 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.159196 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.197441 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 18 00:12:36 crc kubenswrapper[4983]: W0318 00:12:36.202328 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod551c788d_c0f2_4c95_b728_947250e387ef.slice/crio-52e4de99e5c15fa8cde703859bdfa8f0c6064f5584fee52294babadf5db9a4a9 WatchSource:0}: Error finding container 52e4de99e5c15fa8cde703859bdfa8f0c6064f5584fee52294babadf5db9a4a9: Status 404 returned error can't find the container with id 52e4de99e5c15fa8cde703859bdfa8f0c6064f5584fee52294babadf5db9a4a9 Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.320567 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"551c788d-c0f2-4c95-b728-947250e387ef","Type":"ContainerStarted","Data":"52e4de99e5c15fa8cde703859bdfa8f0c6064f5584fee52294babadf5db9a4a9"} Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.323944 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdsf7" event={"ID":"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419","Type":"ContainerStarted","Data":"0c57ae88b210176a3525859ad937dfb7996aec96712aa77054079e0117d2d061"} Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.326043 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" event={"ID":"a213d891-554c-4985-bc09-c03ab2e58198","Type":"ContainerStarted","Data":"60f77ea01264f56d0db96fca4993a25ae217d082c93f31942b8bbffa1070eabe"} Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.326070 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" event={"ID":"a213d891-554c-4985-bc09-c03ab2e58198","Type":"ContainerStarted","Data":"5c6dcf4adaa0457cfad9919575d364a15ac1a8104cd8e631c71a743f8f24536f"} Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.326739 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.337602 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5mgj" event={"ID":"5e139d3f-fd2e-4e39-96b4-eafd340c10a3","Type":"ContainerStarted","Data":"bbcca1d21c22238458b66d7f99ab6280a63f31d8c17f87b8151b6b9144c33a16"} Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.344138 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.344323 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.368028 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tdsf7" podStartSLOduration=8.658401719 podStartE2EDuration="41.368012949s" podCreationTimestamp="2026-03-18 00:11:55 +0000 UTC" firstStartedPulling="2026-03-18 00:12:02.843523638 +0000 UTC m=+240.740898248" lastFinishedPulling="2026-03-18 00:12:35.553134868 +0000 UTC m=+273.450509478" observedRunningTime="2026-03-18 00:12:36.344357932 +0000 UTC m=+274.241732552" watchObservedRunningTime="2026-03-18 00:12:36.368012949 +0000 UTC m=+274.265387559" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.368845 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t5mgj" podStartSLOduration=4.594591027 podStartE2EDuration="37.368839384s" podCreationTimestamp="2026-03-18 00:11:59 +0000 UTC" firstStartedPulling="2026-03-18 00:12:02.884278934 +0000 UTC m=+240.781653544" lastFinishedPulling="2026-03-18 00:12:35.658527291 +0000 UTC m=+273.555901901" observedRunningTime="2026-03-18 00:12:36.366798532 +0000 UTC m=+274.264173152" watchObservedRunningTime="2026-03-18 00:12:36.368839384 +0000 UTC m=+274.266213984" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.370912 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78tfj" event={"ID":"75077192-24a6-4efb-8484-bea2db4a4812","Type":"ContainerStarted","Data":"1019c075b0604a6ba220f5a861fffd2e41468e2909027aeb6459348f11fe7439"} Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.378394 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lqjwd" event={"ID":"5099d24e-b4a3-4990-bc90-980e0188ef3e","Type":"ContainerStarted","Data":"34168a0cd52b8e82778cf0197d2d2f3908e74aea5d3fc4a95194966c904725fb"} Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.388459 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" podStartSLOduration=6.388444038 podStartE2EDuration="6.388444038s" podCreationTimestamp="2026-03-18 00:12:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:12:36.37991953 +0000 UTC m=+274.277294150" watchObservedRunningTime="2026-03-18 00:12:36.388444038 +0000 UTC m=+274.285818648" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.430714 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lqjwd" podStartSLOduration=6.750191646 podStartE2EDuration="39.430699308s" podCreationTimestamp="2026-03-18 00:11:57 +0000 UTC" firstStartedPulling="2026-03-18 00:12:02.895499499 +0000 UTC m=+240.792874109" lastFinishedPulling="2026-03-18 00:12:35.576007161 +0000 UTC m=+273.473381771" observedRunningTime="2026-03-18 00:12:36.408759593 +0000 UTC m=+274.306134203" watchObservedRunningTime="2026-03-18 00:12:36.430699308 +0000 UTC m=+274.328073918" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.431877 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-78tfj" podStartSLOduration=5.6857910799999996 podStartE2EDuration="38.431870594s" podCreationTimestamp="2026-03-18 00:11:58 +0000 UTC" firstStartedPulling="2026-03-18 00:12:02.885247813 +0000 UTC m=+240.782622423" lastFinishedPulling="2026-03-18 00:12:35.631327337 +0000 UTC m=+273.528701937" observedRunningTime="2026-03-18 00:12:36.429959966 +0000 UTC m=+274.327334586" watchObservedRunningTime="2026-03-18 00:12:36.431870594 +0000 UTC m=+274.329245204" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.453049 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.540001 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.540036 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.581461 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:12:36 crc kubenswrapper[4983]: I0318 00:12:36.736078 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:37 crc kubenswrapper[4983]: I0318 00:12:37.200882 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-wnc92" podUID="a81a1b18-1976-45f7-97f5-4dcd7f745af3" containerName="registry-server" probeResult="failure" output=< Mar 18 00:12:37 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Mar 18 00:12:37 crc kubenswrapper[4983]: > Mar 18 00:12:37 crc kubenswrapper[4983]: I0318 00:12:37.383179 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"551c788d-c0f2-4c95-b728-947250e387ef","Type":"ContainerStarted","Data":"816caf1c5136ad61ac657f0783d3f261737a228d3f657dd7503201c784619ab5"} Mar 18 00:12:37 crc kubenswrapper[4983]: I0318 00:12:37.391862 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tdsf7" podUID="b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" containerName="registry-server" probeResult="failure" output=< Mar 18 00:12:37 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Mar 18 00:12:37 crc kubenswrapper[4983]: > Mar 18 00:12:37 crc kubenswrapper[4983]: I0318 00:12:37.404042 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.40401246 podStartE2EDuration="2.40401246s" podCreationTimestamp="2026-03-18 00:12:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:12:37.400906596 +0000 UTC m=+275.298281216" watchObservedRunningTime="2026-03-18 00:12:37.40401246 +0000 UTC m=+275.301387070" Mar 18 00:12:38 crc kubenswrapper[4983]: I0318 00:12:38.136456 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:12:38 crc kubenswrapper[4983]: I0318 00:12:38.136533 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:12:38 crc kubenswrapper[4983]: I0318 00:12:38.530104 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:12:38 crc kubenswrapper[4983]: I0318 00:12:38.530169 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:12:38 crc kubenswrapper[4983]: I0318 00:12:38.567465 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:12:39 crc kubenswrapper[4983]: I0318 00:12:39.150573 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:12:39 crc kubenswrapper[4983]: I0318 00:12:39.150698 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:12:39 crc kubenswrapper[4983]: I0318 00:12:39.184160 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-lqjwd" podUID="5099d24e-b4a3-4990-bc90-980e0188ef3e" containerName="registry-server" probeResult="failure" output=< Mar 18 00:12:39 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Mar 18 00:12:39 crc kubenswrapper[4983]: > Mar 18 00:12:39 crc kubenswrapper[4983]: I0318 00:12:39.435204 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:12:39 crc kubenswrapper[4983]: I0318 00:12:39.541411 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:12:39 crc kubenswrapper[4983]: I0318 00:12:39.541455 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:12:40 crc kubenswrapper[4983]: I0318 00:12:40.189218 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-78tfj" podUID="75077192-24a6-4efb-8484-bea2db4a4812" containerName="registry-server" probeResult="failure" output=< Mar 18 00:12:40 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Mar 18 00:12:40 crc kubenswrapper[4983]: > Mar 18 00:12:40 crc kubenswrapper[4983]: I0318 00:12:40.585143 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t5mgj" podUID="5e139d3f-fd2e-4e39-96b4-eafd340c10a3" containerName="registry-server" probeResult="failure" output=< Mar 18 00:12:40 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Mar 18 00:12:40 crc kubenswrapper[4983]: > Mar 18 00:12:40 crc kubenswrapper[4983]: I0318 00:12:40.602470 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fs978"] Mar 18 00:12:41 crc kubenswrapper[4983]: I0318 00:12:41.404406 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fs978" podUID="089eb8e9-84d9-4320-97ec-9d868b858d3a" containerName="registry-server" containerID="cri-o://d2f46f717106369e8aad5052ffd0791f4b0713776fc894fee3b6827e42ea00a5" gracePeriod=2 Mar 18 00:12:42 crc kubenswrapper[4983]: I0318 00:12:42.412397 4983 generic.go:334] "Generic (PLEG): container finished" podID="089eb8e9-84d9-4320-97ec-9d868b858d3a" containerID="d2f46f717106369e8aad5052ffd0791f4b0713776fc894fee3b6827e42ea00a5" exitCode=0 Mar 18 00:12:42 crc kubenswrapper[4983]: I0318 00:12:42.412451 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fs978" event={"ID":"089eb8e9-84d9-4320-97ec-9d868b858d3a","Type":"ContainerDied","Data":"d2f46f717106369e8aad5052ffd0791f4b0713776fc894fee3b6827e42ea00a5"} Mar 18 00:12:42 crc kubenswrapper[4983]: I0318 00:12:42.660196 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:12:42 crc kubenswrapper[4983]: I0318 00:12:42.763340 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssbd7\" (UniqueName: \"kubernetes.io/projected/089eb8e9-84d9-4320-97ec-9d868b858d3a-kube-api-access-ssbd7\") pod \"089eb8e9-84d9-4320-97ec-9d868b858d3a\" (UID: \"089eb8e9-84d9-4320-97ec-9d868b858d3a\") " Mar 18 00:12:42 crc kubenswrapper[4983]: I0318 00:12:42.763442 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089eb8e9-84d9-4320-97ec-9d868b858d3a-catalog-content\") pod \"089eb8e9-84d9-4320-97ec-9d868b858d3a\" (UID: \"089eb8e9-84d9-4320-97ec-9d868b858d3a\") " Mar 18 00:12:42 crc kubenswrapper[4983]: I0318 00:12:42.763550 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089eb8e9-84d9-4320-97ec-9d868b858d3a-utilities\") pod \"089eb8e9-84d9-4320-97ec-9d868b858d3a\" (UID: \"089eb8e9-84d9-4320-97ec-9d868b858d3a\") " Mar 18 00:12:42 crc kubenswrapper[4983]: I0318 00:12:42.764691 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/089eb8e9-84d9-4320-97ec-9d868b858d3a-utilities" (OuterVolumeSpecName: "utilities") pod "089eb8e9-84d9-4320-97ec-9d868b858d3a" (UID: "089eb8e9-84d9-4320-97ec-9d868b858d3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:12:42 crc kubenswrapper[4983]: I0318 00:12:42.765198 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/089eb8e9-84d9-4320-97ec-9d868b858d3a-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:42 crc kubenswrapper[4983]: I0318 00:12:42.774290 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/089eb8e9-84d9-4320-97ec-9d868b858d3a-kube-api-access-ssbd7" (OuterVolumeSpecName: "kube-api-access-ssbd7") pod "089eb8e9-84d9-4320-97ec-9d868b858d3a" (UID: "089eb8e9-84d9-4320-97ec-9d868b858d3a"). InnerVolumeSpecName "kube-api-access-ssbd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:42 crc kubenswrapper[4983]: I0318 00:12:42.789578 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/089eb8e9-84d9-4320-97ec-9d868b858d3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "089eb8e9-84d9-4320-97ec-9d868b858d3a" (UID: "089eb8e9-84d9-4320-97ec-9d868b858d3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:12:42 crc kubenswrapper[4983]: I0318 00:12:42.867631 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssbd7\" (UniqueName: \"kubernetes.io/projected/089eb8e9-84d9-4320-97ec-9d868b858d3a-kube-api-access-ssbd7\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:42 crc kubenswrapper[4983]: I0318 00:12:42.867908 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/089eb8e9-84d9-4320-97ec-9d868b858d3a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.259755 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.259817 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.259873 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.260559 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062"} pod="openshift-machine-config-operator/machine-config-daemon-jczfs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.260656 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" containerID="cri-o://ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062" gracePeriod=600 Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.421514 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fs978" event={"ID":"089eb8e9-84d9-4320-97ec-9d868b858d3a","Type":"ContainerDied","Data":"ff9d1df6bac87226e656c720843582a61e3dc5b1f4be7103b31cb43709f2f298"} Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.421570 4983 scope.go:117] "RemoveContainer" containerID="d2f46f717106369e8aad5052ffd0791f4b0713776fc894fee3b6827e42ea00a5" Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.421530 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fs978" Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.426484 4983 generic.go:334] "Generic (PLEG): container finished" podID="482e2afe-48be-4ea2-8bda-52849c901b52" containerID="ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062" exitCode=0 Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.426524 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerDied","Data":"ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062"} Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.438261 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fs978"] Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.448723 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fs978"] Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.465803 4983 scope.go:117] "RemoveContainer" containerID="a8c73084999085c4d55387181fe5c17114aa76abc0ff6b47f2d763fbda787e55" Mar 18 00:12:43 crc kubenswrapper[4983]: I0318 00:12:43.494996 4983 scope.go:117] "RemoveContainer" containerID="72fe3e72ff617dd5535918aea93cd17cff83181b2b012c44a2e3d2511ab3189e" Mar 18 00:12:44 crc kubenswrapper[4983]: I0318 00:12:44.436985 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerStarted","Data":"7e5e3b4eea0e0cda0f8f16612d694e8e56064b538ff4f0543482bc9f225a5195"} Mar 18 00:12:44 crc kubenswrapper[4983]: I0318 00:12:44.858869 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="089eb8e9-84d9-4320-97ec-9d868b858d3a" path="/var/lib/kubelet/pods/089eb8e9-84d9-4320-97ec-9d868b858d3a/volumes" Mar 18 00:12:46 crc kubenswrapper[4983]: I0318 00:12:46.219464 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:12:46 crc kubenswrapper[4983]: I0318 00:12:46.280599 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:12:46 crc kubenswrapper[4983]: I0318 00:12:46.397985 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:12:46 crc kubenswrapper[4983]: I0318 00:12:46.460231 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:12:46 crc kubenswrapper[4983]: I0318 00:12:46.577431 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:12:48 crc kubenswrapper[4983]: I0318 00:12:48.210680 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:12:48 crc kubenswrapper[4983]: I0318 00:12:48.284074 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:12:48 crc kubenswrapper[4983]: I0318 00:12:48.612999 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kxm7g"] Mar 18 00:12:48 crc kubenswrapper[4983]: I0318 00:12:48.613392 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kxm7g" podUID="41c73828-774d-4440-a31d-c7cca08b063e" containerName="registry-server" containerID="cri-o://bd9524e52f559d4990bcf31dc6a24772a6d37c53371a731f5bd2501a1892e07e" gracePeriod=2 Mar 18 00:12:48 crc kubenswrapper[4983]: I0318 00:12:48.808399 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tdsf7"] Mar 18 00:12:48 crc kubenswrapper[4983]: I0318 00:12:48.808788 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tdsf7" podUID="b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" containerName="registry-server" containerID="cri-o://0c57ae88b210176a3525859ad937dfb7996aec96712aa77054079e0117d2d061" gracePeriod=2 Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.140560 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.218123 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.261606 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.261905 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c73828-774d-4440-a31d-c7cca08b063e-catalog-content\") pod \"41c73828-774d-4440-a31d-c7cca08b063e\" (UID: \"41c73828-774d-4440-a31d-c7cca08b063e\") " Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.262028 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c73828-774d-4440-a31d-c7cca08b063e-utilities\") pod \"41c73828-774d-4440-a31d-c7cca08b063e\" (UID: \"41c73828-774d-4440-a31d-c7cca08b063e\") " Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.262093 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvvt9\" (UniqueName: \"kubernetes.io/projected/41c73828-774d-4440-a31d-c7cca08b063e-kube-api-access-bvvt9\") pod \"41c73828-774d-4440-a31d-c7cca08b063e\" (UID: \"41c73828-774d-4440-a31d-c7cca08b063e\") " Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.264756 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41c73828-774d-4440-a31d-c7cca08b063e-utilities" (OuterVolumeSpecName: "utilities") pod "41c73828-774d-4440-a31d-c7cca08b063e" (UID: "41c73828-774d-4440-a31d-c7cca08b063e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.271475 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41c73828-774d-4440-a31d-c7cca08b063e-kube-api-access-bvvt9" (OuterVolumeSpecName: "kube-api-access-bvvt9") pod "41c73828-774d-4440-a31d-c7cca08b063e" (UID: "41c73828-774d-4440-a31d-c7cca08b063e"). InnerVolumeSpecName "kube-api-access-bvvt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.271946 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.321322 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41c73828-774d-4440-a31d-c7cca08b063e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41c73828-774d-4440-a31d-c7cca08b063e" (UID: "41c73828-774d-4440-a31d-c7cca08b063e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.363771 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-utilities\") pod \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\" (UID: \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\") " Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.363875 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mz95g\" (UniqueName: \"kubernetes.io/projected/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-kube-api-access-mz95g\") pod \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\" (UID: \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\") " Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.363920 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-catalog-content\") pod \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\" (UID: \"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419\") " Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.364215 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c73828-774d-4440-a31d-c7cca08b063e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.364241 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c73828-774d-4440-a31d-c7cca08b063e-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.364257 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvvt9\" (UniqueName: \"kubernetes.io/projected/41c73828-774d-4440-a31d-c7cca08b063e-kube-api-access-bvvt9\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.364582 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-utilities" (OuterVolumeSpecName: "utilities") pod "b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" (UID: "b2503685-ddb8-4ff3-bdcf-2d7f4d39f419"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.368075 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-kube-api-access-mz95g" (OuterVolumeSpecName: "kube-api-access-mz95g") pod "b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" (UID: "b2503685-ddb8-4ff3-bdcf-2d7f4d39f419"). InnerVolumeSpecName "kube-api-access-mz95g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.432553 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" (UID: "b2503685-ddb8-4ff3-bdcf-2d7f4d39f419"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.465795 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.465834 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mz95g\" (UniqueName: \"kubernetes.io/projected/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-kube-api-access-mz95g\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.465848 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.475687 4983 generic.go:334] "Generic (PLEG): container finished" podID="41c73828-774d-4440-a31d-c7cca08b063e" containerID="bd9524e52f559d4990bcf31dc6a24772a6d37c53371a731f5bd2501a1892e07e" exitCode=0 Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.475818 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxm7g" event={"ID":"41c73828-774d-4440-a31d-c7cca08b063e","Type":"ContainerDied","Data":"bd9524e52f559d4990bcf31dc6a24772a6d37c53371a731f5bd2501a1892e07e"} Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.475748 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxm7g" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.475973 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxm7g" event={"ID":"41c73828-774d-4440-a31d-c7cca08b063e","Type":"ContainerDied","Data":"f22a7aa36053005ecbfcf5f5630ac8e00b25957cb8d9814ad3d44a8a65e8fc14"} Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.476025 4983 scope.go:117] "RemoveContainer" containerID="bd9524e52f559d4990bcf31dc6a24772a6d37c53371a731f5bd2501a1892e07e" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.481237 4983 generic.go:334] "Generic (PLEG): container finished" podID="b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" containerID="0c57ae88b210176a3525859ad937dfb7996aec96712aa77054079e0117d2d061" exitCode=0 Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.481279 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdsf7" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.481353 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdsf7" event={"ID":"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419","Type":"ContainerDied","Data":"0c57ae88b210176a3525859ad937dfb7996aec96712aa77054079e0117d2d061"} Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.481417 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdsf7" event={"ID":"b2503685-ddb8-4ff3-bdcf-2d7f4d39f419","Type":"ContainerDied","Data":"49c28dcce297ab899b65e2b0b8348b7f7d53f5c2ca574c1c69964ccd90e77b90"} Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.513899 4983 scope.go:117] "RemoveContainer" containerID="104f78936c8abfe492c6fa4a7e951f6dde1c4f8d8d79f7f62f1653409d12bf31" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.519179 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tdsf7"] Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.525814 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tdsf7"] Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.546108 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kxm7g"] Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.548360 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kxm7g"] Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.555208 4983 scope.go:117] "RemoveContainer" containerID="8f61cdd4663f4becd3fad3fa742dc3a077ac4cb060a0fd09c831e16cb38bbbbd" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.575107 4983 scope.go:117] "RemoveContainer" containerID="bd9524e52f559d4990bcf31dc6a24772a6d37c53371a731f5bd2501a1892e07e" Mar 18 00:12:49 crc kubenswrapper[4983]: E0318 00:12:49.575671 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd9524e52f559d4990bcf31dc6a24772a6d37c53371a731f5bd2501a1892e07e\": container with ID starting with bd9524e52f559d4990bcf31dc6a24772a6d37c53371a731f5bd2501a1892e07e not found: ID does not exist" containerID="bd9524e52f559d4990bcf31dc6a24772a6d37c53371a731f5bd2501a1892e07e" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.575717 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd9524e52f559d4990bcf31dc6a24772a6d37c53371a731f5bd2501a1892e07e"} err="failed to get container status \"bd9524e52f559d4990bcf31dc6a24772a6d37c53371a731f5bd2501a1892e07e\": rpc error: code = NotFound desc = could not find container \"bd9524e52f559d4990bcf31dc6a24772a6d37c53371a731f5bd2501a1892e07e\": container with ID starting with bd9524e52f559d4990bcf31dc6a24772a6d37c53371a731f5bd2501a1892e07e not found: ID does not exist" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.575749 4983 scope.go:117] "RemoveContainer" containerID="104f78936c8abfe492c6fa4a7e951f6dde1c4f8d8d79f7f62f1653409d12bf31" Mar 18 00:12:49 crc kubenswrapper[4983]: E0318 00:12:49.576417 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"104f78936c8abfe492c6fa4a7e951f6dde1c4f8d8d79f7f62f1653409d12bf31\": container with ID starting with 104f78936c8abfe492c6fa4a7e951f6dde1c4f8d8d79f7f62f1653409d12bf31 not found: ID does not exist" containerID="104f78936c8abfe492c6fa4a7e951f6dde1c4f8d8d79f7f62f1653409d12bf31" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.576512 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"104f78936c8abfe492c6fa4a7e951f6dde1c4f8d8d79f7f62f1653409d12bf31"} err="failed to get container status \"104f78936c8abfe492c6fa4a7e951f6dde1c4f8d8d79f7f62f1653409d12bf31\": rpc error: code = NotFound desc = could not find container \"104f78936c8abfe492c6fa4a7e951f6dde1c4f8d8d79f7f62f1653409d12bf31\": container with ID starting with 104f78936c8abfe492c6fa4a7e951f6dde1c4f8d8d79f7f62f1653409d12bf31 not found: ID does not exist" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.576586 4983 scope.go:117] "RemoveContainer" containerID="8f61cdd4663f4becd3fad3fa742dc3a077ac4cb060a0fd09c831e16cb38bbbbd" Mar 18 00:12:49 crc kubenswrapper[4983]: E0318 00:12:49.577139 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f61cdd4663f4becd3fad3fa742dc3a077ac4cb060a0fd09c831e16cb38bbbbd\": container with ID starting with 8f61cdd4663f4becd3fad3fa742dc3a077ac4cb060a0fd09c831e16cb38bbbbd not found: ID does not exist" containerID="8f61cdd4663f4becd3fad3fa742dc3a077ac4cb060a0fd09c831e16cb38bbbbd" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.577174 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f61cdd4663f4becd3fad3fa742dc3a077ac4cb060a0fd09c831e16cb38bbbbd"} err="failed to get container status \"8f61cdd4663f4becd3fad3fa742dc3a077ac4cb060a0fd09c831e16cb38bbbbd\": rpc error: code = NotFound desc = could not find container \"8f61cdd4663f4becd3fad3fa742dc3a077ac4cb060a0fd09c831e16cb38bbbbd\": container with ID starting with 8f61cdd4663f4becd3fad3fa742dc3a077ac4cb060a0fd09c831e16cb38bbbbd not found: ID does not exist" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.577192 4983 scope.go:117] "RemoveContainer" containerID="0c57ae88b210176a3525859ad937dfb7996aec96712aa77054079e0117d2d061" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.596948 4983 scope.go:117] "RemoveContainer" containerID="684451e4702221db75489b9ec2b8673dc727d8bd912649ca7c5d51787777f1dd" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.599191 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.617250 4983 scope.go:117] "RemoveContainer" containerID="1bd334234e1433a96fc41dbf0a6d47cf493ed4dcb3d95118d38b638aee621620" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.637115 4983 scope.go:117] "RemoveContainer" containerID="0c57ae88b210176a3525859ad937dfb7996aec96712aa77054079e0117d2d061" Mar 18 00:12:49 crc kubenswrapper[4983]: E0318 00:12:49.637709 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c57ae88b210176a3525859ad937dfb7996aec96712aa77054079e0117d2d061\": container with ID starting with 0c57ae88b210176a3525859ad937dfb7996aec96712aa77054079e0117d2d061 not found: ID does not exist" containerID="0c57ae88b210176a3525859ad937dfb7996aec96712aa77054079e0117d2d061" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.637758 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c57ae88b210176a3525859ad937dfb7996aec96712aa77054079e0117d2d061"} err="failed to get container status \"0c57ae88b210176a3525859ad937dfb7996aec96712aa77054079e0117d2d061\": rpc error: code = NotFound desc = could not find container \"0c57ae88b210176a3525859ad937dfb7996aec96712aa77054079e0117d2d061\": container with ID starting with 0c57ae88b210176a3525859ad937dfb7996aec96712aa77054079e0117d2d061 not found: ID does not exist" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.637791 4983 scope.go:117] "RemoveContainer" containerID="684451e4702221db75489b9ec2b8673dc727d8bd912649ca7c5d51787777f1dd" Mar 18 00:12:49 crc kubenswrapper[4983]: E0318 00:12:49.638342 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"684451e4702221db75489b9ec2b8673dc727d8bd912649ca7c5d51787777f1dd\": container with ID starting with 684451e4702221db75489b9ec2b8673dc727d8bd912649ca7c5d51787777f1dd not found: ID does not exist" containerID="684451e4702221db75489b9ec2b8673dc727d8bd912649ca7c5d51787777f1dd" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.638433 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"684451e4702221db75489b9ec2b8673dc727d8bd912649ca7c5d51787777f1dd"} err="failed to get container status \"684451e4702221db75489b9ec2b8673dc727d8bd912649ca7c5d51787777f1dd\": rpc error: code = NotFound desc = could not find container \"684451e4702221db75489b9ec2b8673dc727d8bd912649ca7c5d51787777f1dd\": container with ID starting with 684451e4702221db75489b9ec2b8673dc727d8bd912649ca7c5d51787777f1dd not found: ID does not exist" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.638472 4983 scope.go:117] "RemoveContainer" containerID="1bd334234e1433a96fc41dbf0a6d47cf493ed4dcb3d95118d38b638aee621620" Mar 18 00:12:49 crc kubenswrapper[4983]: E0318 00:12:49.638891 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bd334234e1433a96fc41dbf0a6d47cf493ed4dcb3d95118d38b638aee621620\": container with ID starting with 1bd334234e1433a96fc41dbf0a6d47cf493ed4dcb3d95118d38b638aee621620 not found: ID does not exist" containerID="1bd334234e1433a96fc41dbf0a6d47cf493ed4dcb3d95118d38b638aee621620" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.638974 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bd334234e1433a96fc41dbf0a6d47cf493ed4dcb3d95118d38b638aee621620"} err="failed to get container status \"1bd334234e1433a96fc41dbf0a6d47cf493ed4dcb3d95118d38b638aee621620\": rpc error: code = NotFound desc = could not find container \"1bd334234e1433a96fc41dbf0a6d47cf493ed4dcb3d95118d38b638aee621620\": container with ID starting with 1bd334234e1433a96fc41dbf0a6d47cf493ed4dcb3d95118d38b638aee621620 not found: ID does not exist" Mar 18 00:12:49 crc kubenswrapper[4983]: I0318 00:12:49.651567 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:12:50 crc kubenswrapper[4983]: I0318 00:12:50.856667 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41c73828-774d-4440-a31d-c7cca08b063e" path="/var/lib/kubelet/pods/41c73828-774d-4440-a31d-c7cca08b063e/volumes" Mar 18 00:12:50 crc kubenswrapper[4983]: I0318 00:12:50.859526 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" path="/var/lib/kubelet/pods/b2503685-ddb8-4ff3-bdcf-2d7f4d39f419/volumes" Mar 18 00:12:50 crc kubenswrapper[4983]: I0318 00:12:50.961694 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5b768657c7-9dlrt"] Mar 18 00:12:50 crc kubenswrapper[4983]: I0318 00:12:50.963983 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" podUID="94fb92b4-1a88-43dd-b6b6-0b16e5157a4f" containerName="controller-manager" containerID="cri-o://5f498d046341130439546e497a8bc3606487d4187f16d59dcdc8c2faf2cd4ad0" gracePeriod=30 Mar 18 00:12:50 crc kubenswrapper[4983]: I0318 00:12:50.969848 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv"] Mar 18 00:12:50 crc kubenswrapper[4983]: I0318 00:12:50.970657 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" podUID="a213d891-554c-4985-bc09-c03ab2e58198" containerName="route-controller-manager" containerID="cri-o://60f77ea01264f56d0db96fca4993a25ae217d082c93f31942b8bbffa1070eabe" gracePeriod=30 Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.499651 4983 generic.go:334] "Generic (PLEG): container finished" podID="94fb92b4-1a88-43dd-b6b6-0b16e5157a4f" containerID="5f498d046341130439546e497a8bc3606487d4187f16d59dcdc8c2faf2cd4ad0" exitCode=0 Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.499701 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" event={"ID":"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f","Type":"ContainerDied","Data":"5f498d046341130439546e497a8bc3606487d4187f16d59dcdc8c2faf2cd4ad0"} Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.499861 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.501235 4983 generic.go:334] "Generic (PLEG): container finished" podID="a213d891-554c-4985-bc09-c03ab2e58198" containerID="60f77ea01264f56d0db96fca4993a25ae217d082c93f31942b8bbffa1070eabe" exitCode=0 Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.501271 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" event={"ID":"a213d891-554c-4985-bc09-c03ab2e58198","Type":"ContainerDied","Data":"60f77ea01264f56d0db96fca4993a25ae217d082c93f31942b8bbffa1070eabe"} Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.501339 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" event={"ID":"a213d891-554c-4985-bc09-c03ab2e58198","Type":"ContainerDied","Data":"5c6dcf4adaa0457cfad9919575d364a15ac1a8104cd8e631c71a743f8f24536f"} Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.501362 4983 scope.go:117] "RemoveContainer" containerID="60f77ea01264f56d0db96fca4993a25ae217d082c93f31942b8bbffa1070eabe" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.517972 4983 scope.go:117] "RemoveContainer" containerID="60f77ea01264f56d0db96fca4993a25ae217d082c93f31942b8bbffa1070eabe" Mar 18 00:12:51 crc kubenswrapper[4983]: E0318 00:12:51.518455 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60f77ea01264f56d0db96fca4993a25ae217d082c93f31942b8bbffa1070eabe\": container with ID starting with 60f77ea01264f56d0db96fca4993a25ae217d082c93f31942b8bbffa1070eabe not found: ID does not exist" containerID="60f77ea01264f56d0db96fca4993a25ae217d082c93f31942b8bbffa1070eabe" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.518499 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60f77ea01264f56d0db96fca4993a25ae217d082c93f31942b8bbffa1070eabe"} err="failed to get container status \"60f77ea01264f56d0db96fca4993a25ae217d082c93f31942b8bbffa1070eabe\": rpc error: code = NotFound desc = could not find container \"60f77ea01264f56d0db96fca4993a25ae217d082c93f31942b8bbffa1070eabe\": container with ID starting with 60f77ea01264f56d0db96fca4993a25ae217d082c93f31942b8bbffa1070eabe not found: ID does not exist" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.560491 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.605459 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a213d891-554c-4985-bc09-c03ab2e58198-serving-cert\") pod \"a213d891-554c-4985-bc09-c03ab2e58198\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.605531 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptkns\" (UniqueName: \"kubernetes.io/projected/a213d891-554c-4985-bc09-c03ab2e58198-kube-api-access-ptkns\") pod \"a213d891-554c-4985-bc09-c03ab2e58198\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.605561 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a213d891-554c-4985-bc09-c03ab2e58198-client-ca\") pod \"a213d891-554c-4985-bc09-c03ab2e58198\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.605595 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a213d891-554c-4985-bc09-c03ab2e58198-config\") pod \"a213d891-554c-4985-bc09-c03ab2e58198\" (UID: \"a213d891-554c-4985-bc09-c03ab2e58198\") " Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.606491 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a213d891-554c-4985-bc09-c03ab2e58198-client-ca" (OuterVolumeSpecName: "client-ca") pod "a213d891-554c-4985-bc09-c03ab2e58198" (UID: "a213d891-554c-4985-bc09-c03ab2e58198"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.606691 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a213d891-554c-4985-bc09-c03ab2e58198-config" (OuterVolumeSpecName: "config") pod "a213d891-554c-4985-bc09-c03ab2e58198" (UID: "a213d891-554c-4985-bc09-c03ab2e58198"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.610215 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a213d891-554c-4985-bc09-c03ab2e58198-kube-api-access-ptkns" (OuterVolumeSpecName: "kube-api-access-ptkns") pod "a213d891-554c-4985-bc09-c03ab2e58198" (UID: "a213d891-554c-4985-bc09-c03ab2e58198"). InnerVolumeSpecName "kube-api-access-ptkns". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.611157 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a213d891-554c-4985-bc09-c03ab2e58198-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a213d891-554c-4985-bc09-c03ab2e58198" (UID: "a213d891-554c-4985-bc09-c03ab2e58198"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.707143 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-proxy-ca-bundles\") pod \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.707198 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-client-ca\") pod \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.707242 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-config\") pod \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.707273 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2b8r\" (UniqueName: \"kubernetes.io/projected/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-kube-api-access-w2b8r\") pod \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.707316 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-serving-cert\") pod \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\" (UID: \"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f\") " Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.707511 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptkns\" (UniqueName: \"kubernetes.io/projected/a213d891-554c-4985-bc09-c03ab2e58198-kube-api-access-ptkns\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.707525 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a213d891-554c-4985-bc09-c03ab2e58198-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.707534 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a213d891-554c-4985-bc09-c03ab2e58198-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.707568 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a213d891-554c-4985-bc09-c03ab2e58198-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.708019 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "94fb92b4-1a88-43dd-b6b6-0b16e5157a4f" (UID: "94fb92b4-1a88-43dd-b6b6-0b16e5157a4f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.708028 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-client-ca" (OuterVolumeSpecName: "client-ca") pod "94fb92b4-1a88-43dd-b6b6-0b16e5157a4f" (UID: "94fb92b4-1a88-43dd-b6b6-0b16e5157a4f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.708800 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-config" (OuterVolumeSpecName: "config") pod "94fb92b4-1a88-43dd-b6b6-0b16e5157a4f" (UID: "94fb92b4-1a88-43dd-b6b6-0b16e5157a4f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.710327 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "94fb92b4-1a88-43dd-b6b6-0b16e5157a4f" (UID: "94fb92b4-1a88-43dd-b6b6-0b16e5157a4f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.710813 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-kube-api-access-w2b8r" (OuterVolumeSpecName: "kube-api-access-w2b8r") pod "94fb92b4-1a88-43dd-b6b6-0b16e5157a4f" (UID: "94fb92b4-1a88-43dd-b6b6-0b16e5157a4f"). InnerVolumeSpecName "kube-api-access-w2b8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.808678 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.808709 4983 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.808719 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.808728 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:51 crc kubenswrapper[4983]: I0318 00:12:51.808736 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2b8r\" (UniqueName: \"kubernetes.io/projected/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f-kube-api-access-w2b8r\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.240175 4983 patch_prober.go:28] interesting pod/controller-manager-5b768657c7-9dlrt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.63:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.240266 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" podUID="94fb92b4-1a88-43dd-b6b6-0b16e5157a4f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.63:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.522408 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.522405 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b768657c7-9dlrt" event={"ID":"94fb92b4-1a88-43dd-b6b6-0b16e5157a4f","Type":"ContainerDied","Data":"f5ceab20ccd608a5ac263b84b4f311277d5bd49d71b0a0536143184d96cb10e9"} Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.522908 4983 scope.go:117] "RemoveContainer" containerID="5f498d046341130439546e497a8bc3606487d4187f16d59dcdc8c2faf2cd4ad0" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.524032 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.572521 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5b768657c7-9dlrt"] Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.580045 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5b768657c7-9dlrt"] Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.585282 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv"] Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.590186 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b4dc48cdc-v45kv"] Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.759982 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" podUID="66e8121d-9422-4b23-93e9-9f4d350c149f" containerName="oauth-openshift" containerID="cri-o://8b0c8721863a3abbba8bdda49afb6a1e93d345e282e9df701840572628a7871d" gracePeriod=15 Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.852983 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94fb92b4-1a88-43dd-b6b6-0b16e5157a4f" path="/var/lib/kubelet/pods/94fb92b4-1a88-43dd-b6b6-0b16e5157a4f/volumes" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.853867 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a213d891-554c-4985-bc09-c03ab2e58198" path="/var/lib/kubelet/pods/a213d891-554c-4985-bc09-c03ab2e58198/volumes" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.886344 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c"] Mar 18 00:12:52 crc kubenswrapper[4983]: E0318 00:12:52.886649 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089eb8e9-84d9-4320-97ec-9d868b858d3a" containerName="extract-utilities" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.886670 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="089eb8e9-84d9-4320-97ec-9d868b858d3a" containerName="extract-utilities" Mar 18 00:12:52 crc kubenswrapper[4983]: E0318 00:12:52.886687 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089eb8e9-84d9-4320-97ec-9d868b858d3a" containerName="registry-server" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.886698 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="089eb8e9-84d9-4320-97ec-9d868b858d3a" containerName="registry-server" Mar 18 00:12:52 crc kubenswrapper[4983]: E0318 00:12:52.886715 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" containerName="extract-content" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.886723 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" containerName="extract-content" Mar 18 00:12:52 crc kubenswrapper[4983]: E0318 00:12:52.886738 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089eb8e9-84d9-4320-97ec-9d868b858d3a" containerName="extract-content" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.886745 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="089eb8e9-84d9-4320-97ec-9d868b858d3a" containerName="extract-content" Mar 18 00:12:52 crc kubenswrapper[4983]: E0318 00:12:52.886759 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c73828-774d-4440-a31d-c7cca08b063e" containerName="extract-content" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.886767 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c73828-774d-4440-a31d-c7cca08b063e" containerName="extract-content" Mar 18 00:12:52 crc kubenswrapper[4983]: E0318 00:12:52.886779 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c73828-774d-4440-a31d-c7cca08b063e" containerName="extract-utilities" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.886787 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c73828-774d-4440-a31d-c7cca08b063e" containerName="extract-utilities" Mar 18 00:12:52 crc kubenswrapper[4983]: E0318 00:12:52.886796 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94fb92b4-1a88-43dd-b6b6-0b16e5157a4f" containerName="controller-manager" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.886803 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="94fb92b4-1a88-43dd-b6b6-0b16e5157a4f" containerName="controller-manager" Mar 18 00:12:52 crc kubenswrapper[4983]: E0318 00:12:52.886814 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c73828-774d-4440-a31d-c7cca08b063e" containerName="registry-server" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.886822 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c73828-774d-4440-a31d-c7cca08b063e" containerName="registry-server" Mar 18 00:12:52 crc kubenswrapper[4983]: E0318 00:12:52.886837 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" containerName="registry-server" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.886848 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" containerName="registry-server" Mar 18 00:12:52 crc kubenswrapper[4983]: E0318 00:12:52.886862 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" containerName="extract-utilities" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.886871 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" containerName="extract-utilities" Mar 18 00:12:52 crc kubenswrapper[4983]: E0318 00:12:52.886880 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a213d891-554c-4985-bc09-c03ab2e58198" containerName="route-controller-manager" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.886889 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a213d891-554c-4985-bc09-c03ab2e58198" containerName="route-controller-manager" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.887069 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="a213d891-554c-4985-bc09-c03ab2e58198" containerName="route-controller-manager" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.887084 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2503685-ddb8-4ff3-bdcf-2d7f4d39f419" containerName="registry-server" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.887098 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="089eb8e9-84d9-4320-97ec-9d868b858d3a" containerName="registry-server" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.887109 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="41c73828-774d-4440-a31d-c7cca08b063e" containerName="registry-server" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.887122 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="94fb92b4-1a88-43dd-b6b6-0b16e5157a4f" containerName="controller-manager" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.887750 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.891822 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.891995 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-577468b5b7-r4qdg"] Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.892309 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.893673 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.893971 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.894225 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.894477 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.903624 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.905506 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.907717 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c"] Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.909441 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.909749 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.910307 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.910479 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.910625 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.911180 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-577468b5b7-r4qdg"] Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.914186 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.937589 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-config\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.937653 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6vpn\" (UniqueName: \"kubernetes.io/projected/76b2bf27-1695-4652-8469-1b3e34652b5a-kube-api-access-b6vpn\") pod \"route-controller-manager-5bbbd659c4-56s2c\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.937678 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-serving-cert\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.937702 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57bjr\" (UniqueName: \"kubernetes.io/projected/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-kube-api-access-57bjr\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.937731 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/76b2bf27-1695-4652-8469-1b3e34652b5a-client-ca\") pod \"route-controller-manager-5bbbd659c4-56s2c\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.937753 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-proxy-ca-bundles\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.937780 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76b2bf27-1695-4652-8469-1b3e34652b5a-serving-cert\") pod \"route-controller-manager-5bbbd659c4-56s2c\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.937823 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-client-ca\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:52 crc kubenswrapper[4983]: I0318 00:12:52.937848 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76b2bf27-1695-4652-8469-1b3e34652b5a-config\") pod \"route-controller-manager-5bbbd659c4-56s2c\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.045542 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6vpn\" (UniqueName: \"kubernetes.io/projected/76b2bf27-1695-4652-8469-1b3e34652b5a-kube-api-access-b6vpn\") pod \"route-controller-manager-5bbbd659c4-56s2c\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.045598 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-serving-cert\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.045624 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57bjr\" (UniqueName: \"kubernetes.io/projected/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-kube-api-access-57bjr\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.045660 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/76b2bf27-1695-4652-8469-1b3e34652b5a-client-ca\") pod \"route-controller-manager-5bbbd659c4-56s2c\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.045679 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-proxy-ca-bundles\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.045709 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76b2bf27-1695-4652-8469-1b3e34652b5a-serving-cert\") pod \"route-controller-manager-5bbbd659c4-56s2c\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.045771 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-client-ca\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.045793 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76b2bf27-1695-4652-8469-1b3e34652b5a-config\") pod \"route-controller-manager-5bbbd659c4-56s2c\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.045862 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-config\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.046986 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-client-ca\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.047079 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/76b2bf27-1695-4652-8469-1b3e34652b5a-client-ca\") pod \"route-controller-manager-5bbbd659c4-56s2c\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.047648 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-config\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.048109 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-proxy-ca-bundles\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.048298 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76b2bf27-1695-4652-8469-1b3e34652b5a-config\") pod \"route-controller-manager-5bbbd659c4-56s2c\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.068676 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76b2bf27-1695-4652-8469-1b3e34652b5a-serving-cert\") pod \"route-controller-manager-5bbbd659c4-56s2c\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.073818 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-serving-cert\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.075564 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6vpn\" (UniqueName: \"kubernetes.io/projected/76b2bf27-1695-4652-8469-1b3e34652b5a-kube-api-access-b6vpn\") pod \"route-controller-manager-5bbbd659c4-56s2c\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.076452 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57bjr\" (UniqueName: \"kubernetes.io/projected/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-kube-api-access-57bjr\") pod \"controller-manager-577468b5b7-r4qdg\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.169584 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.209430 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t5mgj"] Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.210212 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t5mgj" podUID="5e139d3f-fd2e-4e39-96b4-eafd340c10a3" containerName="registry-server" containerID="cri-o://bbcca1d21c22238458b66d7f99ab6280a63f31d8c17f87b8151b6b9144c33a16" gracePeriod=2 Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.253195 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.277288 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353354 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66e8121d-9422-4b23-93e9-9f4d350c149f-audit-dir\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353429 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-audit-policies\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353468 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-login\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353513 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-router-certs\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353543 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-idp-0-file-data\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353543 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66e8121d-9422-4b23-93e9-9f4d350c149f-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353581 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-service-ca\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353627 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-ocp-branding-template\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353669 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-session\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353725 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-trusted-ca-bundle\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353766 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2p82\" (UniqueName: \"kubernetes.io/projected/66e8121d-9422-4b23-93e9-9f4d350c149f-kube-api-access-m2p82\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353794 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-serving-cert\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353826 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-error\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.353875 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-cliconfig\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.354028 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-provider-selection\") pod \"66e8121d-9422-4b23-93e9-9f4d350c149f\" (UID: \"66e8121d-9422-4b23-93e9-9f4d350c149f\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.354300 4983 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66e8121d-9422-4b23-93e9-9f4d350c149f-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.354508 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.355327 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.357968 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.360126 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.362112 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.362464 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.363517 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.364587 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66e8121d-9422-4b23-93e9-9f4d350c149f-kube-api-access-m2p82" (OuterVolumeSpecName: "kube-api-access-m2p82") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "kube-api-access-m2p82". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.365988 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.366847 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.367473 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.369156 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.370503 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "66e8121d-9422-4b23-93e9-9f4d350c149f" (UID: "66e8121d-9422-4b23-93e9-9f4d350c149f"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.455344 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.455379 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.455397 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.455407 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.455430 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.455440 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.455450 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2p82\" (UniqueName: \"kubernetes.io/projected/66e8121d-9422-4b23-93e9-9f4d350c149f-kube-api-access-m2p82\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.455459 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.455467 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.455475 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.455484 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.455495 4983 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66e8121d-9422-4b23-93e9-9f4d350c149f-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.455503 4983 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66e8121d-9422-4b23-93e9-9f4d350c149f-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.531994 4983 generic.go:334] "Generic (PLEG): container finished" podID="5e139d3f-fd2e-4e39-96b4-eafd340c10a3" containerID="bbcca1d21c22238458b66d7f99ab6280a63f31d8c17f87b8151b6b9144c33a16" exitCode=0 Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.532065 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5mgj" event={"ID":"5e139d3f-fd2e-4e39-96b4-eafd340c10a3","Type":"ContainerDied","Data":"bbcca1d21c22238458b66d7f99ab6280a63f31d8c17f87b8151b6b9144c33a16"} Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.533734 4983 generic.go:334] "Generic (PLEG): container finished" podID="66e8121d-9422-4b23-93e9-9f4d350c149f" containerID="8b0c8721863a3abbba8bdda49afb6a1e93d345e282e9df701840572628a7871d" exitCode=0 Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.533785 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.533811 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" event={"ID":"66e8121d-9422-4b23-93e9-9f4d350c149f","Type":"ContainerDied","Data":"8b0c8721863a3abbba8bdda49afb6a1e93d345e282e9df701840572628a7871d"} Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.533833 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6t6lv" event={"ID":"66e8121d-9422-4b23-93e9-9f4d350c149f","Type":"ContainerDied","Data":"cc53b56e437c38c8592783cbafa55e920436da4a14527406ce8715bbf6dda778"} Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.533848 4983 scope.go:117] "RemoveContainer" containerID="8b0c8721863a3abbba8bdda49afb6a1e93d345e282e9df701840572628a7871d" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.553780 4983 scope.go:117] "RemoveContainer" containerID="8b0c8721863a3abbba8bdda49afb6a1e93d345e282e9df701840572628a7871d" Mar 18 00:12:53 crc kubenswrapper[4983]: E0318 00:12:53.555363 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b0c8721863a3abbba8bdda49afb6a1e93d345e282e9df701840572628a7871d\": container with ID starting with 8b0c8721863a3abbba8bdda49afb6a1e93d345e282e9df701840572628a7871d not found: ID does not exist" containerID="8b0c8721863a3abbba8bdda49afb6a1e93d345e282e9df701840572628a7871d" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.555406 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b0c8721863a3abbba8bdda49afb6a1e93d345e282e9df701840572628a7871d"} err="failed to get container status \"8b0c8721863a3abbba8bdda49afb6a1e93d345e282e9df701840572628a7871d\": rpc error: code = NotFound desc = could not find container \"8b0c8721863a3abbba8bdda49afb6a1e93d345e282e9df701840572628a7871d\": container with ID starting with 8b0c8721863a3abbba8bdda49afb6a1e93d345e282e9df701840572628a7871d not found: ID does not exist" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.570323 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6t6lv"] Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.573244 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6t6lv"] Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.597748 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.618000 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-577468b5b7-r4qdg"] Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.681603 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c"] Mar 18 00:12:53 crc kubenswrapper[4983]: W0318 00:12:53.694810 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76b2bf27_1695_4652_8469_1b3e34652b5a.slice/crio-63cf13a07e4eb604a0469b39252e6e766d446801cdf8d272b8944fae69dc5d5f WatchSource:0}: Error finding container 63cf13a07e4eb604a0469b39252e6e766d446801cdf8d272b8944fae69dc5d5f: Status 404 returned error can't find the container with id 63cf13a07e4eb604a0469b39252e6e766d446801cdf8d272b8944fae69dc5d5f Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.758315 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-utilities\") pod \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\" (UID: \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.758370 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-catalog-content\") pod \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\" (UID: \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.758482 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72btz\" (UniqueName: \"kubernetes.io/projected/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-kube-api-access-72btz\") pod \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\" (UID: \"5e139d3f-fd2e-4e39-96b4-eafd340c10a3\") " Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.759199 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-utilities" (OuterVolumeSpecName: "utilities") pod "5e139d3f-fd2e-4e39-96b4-eafd340c10a3" (UID: "5e139d3f-fd2e-4e39-96b4-eafd340c10a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.764695 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-kube-api-access-72btz" (OuterVolumeSpecName: "kube-api-access-72btz") pod "5e139d3f-fd2e-4e39-96b4-eafd340c10a3" (UID: "5e139d3f-fd2e-4e39-96b4-eafd340c10a3"). InnerVolumeSpecName "kube-api-access-72btz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.860121 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72btz\" (UniqueName: \"kubernetes.io/projected/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-kube-api-access-72btz\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.860156 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.907331 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e139d3f-fd2e-4e39-96b4-eafd340c10a3" (UID: "5e139d3f-fd2e-4e39-96b4-eafd340c10a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:12:53 crc kubenswrapper[4983]: I0318 00:12:53.960986 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e139d3f-fd2e-4e39-96b4-eafd340c10a3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.544213 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5mgj" event={"ID":"5e139d3f-fd2e-4e39-96b4-eafd340c10a3","Type":"ContainerDied","Data":"6b6fea9a4ccf1425c159036414b60551fc1946ace7ba4bc1f3ee2abfc8050b25"} Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.544269 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5mgj" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.544708 4983 scope.go:117] "RemoveContainer" containerID="bbcca1d21c22238458b66d7f99ab6280a63f31d8c17f87b8151b6b9144c33a16" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.549210 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" event={"ID":"76b2bf27-1695-4652-8469-1b3e34652b5a","Type":"ContainerStarted","Data":"67efbd75adc302d24ec3d6ad67ec126f1fcb357d704f735971cc71b569bee2c4"} Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.549249 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" event={"ID":"76b2bf27-1695-4652-8469-1b3e34652b5a","Type":"ContainerStarted","Data":"63cf13a07e4eb604a0469b39252e6e766d446801cdf8d272b8944fae69dc5d5f"} Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.549441 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.551066 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" event={"ID":"84a117a5-7bb6-499f-8b7f-c9f007b86c1d","Type":"ContainerStarted","Data":"397d9c5429d1a190839b1b3e27106f17ab442797d9881f2ca9ccfd32584458ec"} Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.551096 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" event={"ID":"84a117a5-7bb6-499f-8b7f-c9f007b86c1d","Type":"ContainerStarted","Data":"e05a2c8ce768f95bcd71008cd3e2c91ce02bcbe00ced3e9ff212276c12259417"} Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.551408 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.554898 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.557257 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.566801 4983 scope.go:117] "RemoveContainer" containerID="d9dfb1511dc0a58eacc6af7b10f950487159392752cdfd91ae4dffb5530c4b59" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.581181 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" podStartSLOduration=4.581119302 podStartE2EDuration="4.581119302s" podCreationTimestamp="2026-03-18 00:12:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:12:54.572226632 +0000 UTC m=+292.469601272" watchObservedRunningTime="2026-03-18 00:12:54.581119302 +0000 UTC m=+292.478493912" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.585757 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t5mgj"] Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.592989 4983 scope.go:117] "RemoveContainer" containerID="453aca7b0fae9e0bc9b795bf327ca036152e6e08018a1b820f2bbc94b6f83332" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.592996 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t5mgj"] Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.609753 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" podStartSLOduration=4.609735809 podStartE2EDuration="4.609735809s" podCreationTimestamp="2026-03-18 00:12:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:12:54.606031177 +0000 UTC m=+292.503405807" watchObservedRunningTime="2026-03-18 00:12:54.609735809 +0000 UTC m=+292.507110419" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.855117 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e139d3f-fd2e-4e39-96b4-eafd340c10a3" path="/var/lib/kubelet/pods/5e139d3f-fd2e-4e39-96b4-eafd340c10a3/volumes" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.855982 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66e8121d-9422-4b23-93e9-9f4d350c149f" path="/var/lib/kubelet/pods/66e8121d-9422-4b23-93e9-9f4d350c149f/volumes" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.882746 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk"] Mar 18 00:12:54 crc kubenswrapper[4983]: E0318 00:12:54.882984 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66e8121d-9422-4b23-93e9-9f4d350c149f" containerName="oauth-openshift" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.882998 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="66e8121d-9422-4b23-93e9-9f4d350c149f" containerName="oauth-openshift" Mar 18 00:12:54 crc kubenswrapper[4983]: E0318 00:12:54.883005 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e139d3f-fd2e-4e39-96b4-eafd340c10a3" containerName="extract-content" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.883010 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e139d3f-fd2e-4e39-96b4-eafd340c10a3" containerName="extract-content" Mar 18 00:12:54 crc kubenswrapper[4983]: E0318 00:12:54.883023 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e139d3f-fd2e-4e39-96b4-eafd340c10a3" containerName="registry-server" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.883029 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e139d3f-fd2e-4e39-96b4-eafd340c10a3" containerName="registry-server" Mar 18 00:12:54 crc kubenswrapper[4983]: E0318 00:12:54.883042 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e139d3f-fd2e-4e39-96b4-eafd340c10a3" containerName="extract-utilities" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.883050 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e139d3f-fd2e-4e39-96b4-eafd340c10a3" containerName="extract-utilities" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.883246 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="66e8121d-9422-4b23-93e9-9f4d350c149f" containerName="oauth-openshift" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.883262 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e139d3f-fd2e-4e39-96b4-eafd340c10a3" containerName="registry-server" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.883758 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.888546 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.888606 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.888668 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.888871 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.889117 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.889193 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.890086 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.890132 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.890213 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.890390 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.890611 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.890619 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.928894 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.931513 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.935463 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk"] Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.939311 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.973765 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/93d5bb46-82a3-4e71-a394-bb63d57779de-audit-policies\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.973871 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-user-template-error\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.973906 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.973960 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-service-ca\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.973982 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-user-template-login\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.974009 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wn9s\" (UniqueName: \"kubernetes.io/projected/93d5bb46-82a3-4e71-a394-bb63d57779de-kube-api-access-9wn9s\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.974068 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.974119 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.974158 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-cliconfig\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.974187 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.974212 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/93d5bb46-82a3-4e71-a394-bb63d57779de-audit-dir\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.974271 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-serving-cert\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.974297 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-router-certs\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:54 crc kubenswrapper[4983]: I0318 00:12:54.974374 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-session\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.075419 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-serving-cert\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.075735 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-router-certs\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.075982 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-session\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.076147 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/93d5bb46-82a3-4e71-a394-bb63d57779de-audit-policies\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.076279 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-user-template-error\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.076413 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.076603 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-user-template-login\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.076762 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-service-ca\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.076903 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wn9s\" (UniqueName: \"kubernetes.io/projected/93d5bb46-82a3-4e71-a394-bb63d57779de-kube-api-access-9wn9s\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.077090 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.077225 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.077341 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/93d5bb46-82a3-4e71-a394-bb63d57779de-audit-policies\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.077351 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-cliconfig\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.077447 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.077488 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/93d5bb46-82a3-4e71-a394-bb63d57779de-audit-dir\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.077680 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/93d5bb46-82a3-4e71-a394-bb63d57779de-audit-dir\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.077843 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-service-ca\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.078540 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-cliconfig\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.079554 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.083188 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.083271 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-router-certs\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.083512 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-user-template-error\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.083648 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-session\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.083740 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-serving-cert\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.084841 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.086983 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-user-template-login\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.087094 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/93d5bb46-82a3-4e71-a394-bb63d57779de-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.098097 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wn9s\" (UniqueName: \"kubernetes.io/projected/93d5bb46-82a3-4e71-a394-bb63d57779de-kube-api-access-9wn9s\") pod \"oauth-openshift-85df6bd6d5-k2vgk\" (UID: \"93d5bb46-82a3-4e71-a394-bb63d57779de\") " pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.225349 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:55 crc kubenswrapper[4983]: I0318 00:12:55.744205 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk"] Mar 18 00:12:55 crc kubenswrapper[4983]: W0318 00:12:55.752141 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93d5bb46_82a3_4e71_a394_bb63d57779de.slice/crio-35e2777576e77a5d2c6de647998cd95294d1442a2fcd1584408212c6c6b1b162 WatchSource:0}: Error finding container 35e2777576e77a5d2c6de647998cd95294d1442a2fcd1584408212c6c6b1b162: Status 404 returned error can't find the container with id 35e2777576e77a5d2c6de647998cd95294d1442a2fcd1584408212c6c6b1b162 Mar 18 00:12:56 crc kubenswrapper[4983]: I0318 00:12:56.575380 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" event={"ID":"93d5bb46-82a3-4e71-a394-bb63d57779de","Type":"ContainerStarted","Data":"4eb93e3ca38bf6eadca3da90d4e22c18766f854735f255b609daca9c3c3ab19c"} Mar 18 00:12:56 crc kubenswrapper[4983]: I0318 00:12:56.575801 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" event={"ID":"93d5bb46-82a3-4e71-a394-bb63d57779de","Type":"ContainerStarted","Data":"35e2777576e77a5d2c6de647998cd95294d1442a2fcd1584408212c6c6b1b162"} Mar 18 00:12:56 crc kubenswrapper[4983]: I0318 00:12:56.608361 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" podStartSLOduration=29.608327136 podStartE2EDuration="29.608327136s" podCreationTimestamp="2026-03-18 00:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:12:56.605236372 +0000 UTC m=+294.502611082" watchObservedRunningTime="2026-03-18 00:12:56.608327136 +0000 UTC m=+294.505701796" Mar 18 00:12:57 crc kubenswrapper[4983]: I0318 00:12:57.583567 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:12:57 crc kubenswrapper[4983]: I0318 00:12:57.596384 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-85df6bd6d5-k2vgk" Mar 18 00:13:10 crc kubenswrapper[4983]: I0318 00:13:10.894545 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-577468b5b7-r4qdg"] Mar 18 00:13:10 crc kubenswrapper[4983]: I0318 00:13:10.895534 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" podUID="84a117a5-7bb6-499f-8b7f-c9f007b86c1d" containerName="controller-manager" containerID="cri-o://397d9c5429d1a190839b1b3e27106f17ab442797d9881f2ca9ccfd32584458ec" gracePeriod=30 Mar 18 00:13:10 crc kubenswrapper[4983]: I0318 00:13:10.990190 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c"] Mar 18 00:13:10 crc kubenswrapper[4983]: I0318 00:13:10.990435 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" podUID="76b2bf27-1695-4652-8469-1b3e34652b5a" containerName="route-controller-manager" containerID="cri-o://67efbd75adc302d24ec3d6ad67ec126f1fcb357d704f735971cc71b569bee2c4" gracePeriod=30 Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.445162 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.508650 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76b2bf27-1695-4652-8469-1b3e34652b5a-config\") pod \"76b2bf27-1695-4652-8469-1b3e34652b5a\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.508753 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76b2bf27-1695-4652-8469-1b3e34652b5a-serving-cert\") pod \"76b2bf27-1695-4652-8469-1b3e34652b5a\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.508838 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/76b2bf27-1695-4652-8469-1b3e34652b5a-client-ca\") pod \"76b2bf27-1695-4652-8469-1b3e34652b5a\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.508868 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6vpn\" (UniqueName: \"kubernetes.io/projected/76b2bf27-1695-4652-8469-1b3e34652b5a-kube-api-access-b6vpn\") pod \"76b2bf27-1695-4652-8469-1b3e34652b5a\" (UID: \"76b2bf27-1695-4652-8469-1b3e34652b5a\") " Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.509739 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76b2bf27-1695-4652-8469-1b3e34652b5a-client-ca" (OuterVolumeSpecName: "client-ca") pod "76b2bf27-1695-4652-8469-1b3e34652b5a" (UID: "76b2bf27-1695-4652-8469-1b3e34652b5a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.509776 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76b2bf27-1695-4652-8469-1b3e34652b5a-config" (OuterVolumeSpecName: "config") pod "76b2bf27-1695-4652-8469-1b3e34652b5a" (UID: "76b2bf27-1695-4652-8469-1b3e34652b5a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.515382 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76b2bf27-1695-4652-8469-1b3e34652b5a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "76b2bf27-1695-4652-8469-1b3e34652b5a" (UID: "76b2bf27-1695-4652-8469-1b3e34652b5a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.515837 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76b2bf27-1695-4652-8469-1b3e34652b5a-kube-api-access-b6vpn" (OuterVolumeSpecName: "kube-api-access-b6vpn") pod "76b2bf27-1695-4652-8469-1b3e34652b5a" (UID: "76b2bf27-1695-4652-8469-1b3e34652b5a"). InnerVolumeSpecName "kube-api-access-b6vpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.526588 4983 generic.go:334] "Generic (PLEG): container finished" podID="76b2bf27-1695-4652-8469-1b3e34652b5a" containerID="67efbd75adc302d24ec3d6ad67ec126f1fcb357d704f735971cc71b569bee2c4" exitCode=0 Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.526647 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" event={"ID":"76b2bf27-1695-4652-8469-1b3e34652b5a","Type":"ContainerDied","Data":"67efbd75adc302d24ec3d6ad67ec126f1fcb357d704f735971cc71b569bee2c4"} Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.526671 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" event={"ID":"76b2bf27-1695-4652-8469-1b3e34652b5a","Type":"ContainerDied","Data":"63cf13a07e4eb604a0469b39252e6e766d446801cdf8d272b8944fae69dc5d5f"} Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.526687 4983 scope.go:117] "RemoveContainer" containerID="67efbd75adc302d24ec3d6ad67ec126f1fcb357d704f735971cc71b569bee2c4" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.526770 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.531966 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" event={"ID":"84a117a5-7bb6-499f-8b7f-c9f007b86c1d","Type":"ContainerDied","Data":"397d9c5429d1a190839b1b3e27106f17ab442797d9881f2ca9ccfd32584458ec"} Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.532000 4983 generic.go:334] "Generic (PLEG): container finished" podID="84a117a5-7bb6-499f-8b7f-c9f007b86c1d" containerID="397d9c5429d1a190839b1b3e27106f17ab442797d9881f2ca9ccfd32584458ec" exitCode=0 Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.570957 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.572420 4983 scope.go:117] "RemoveContainer" containerID="67efbd75adc302d24ec3d6ad67ec126f1fcb357d704f735971cc71b569bee2c4" Mar 18 00:13:11 crc kubenswrapper[4983]: E0318 00:13:11.572687 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67efbd75adc302d24ec3d6ad67ec126f1fcb357d704f735971cc71b569bee2c4\": container with ID starting with 67efbd75adc302d24ec3d6ad67ec126f1fcb357d704f735971cc71b569bee2c4 not found: ID does not exist" containerID="67efbd75adc302d24ec3d6ad67ec126f1fcb357d704f735971cc71b569bee2c4" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.572826 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67efbd75adc302d24ec3d6ad67ec126f1fcb357d704f735971cc71b569bee2c4"} err="failed to get container status \"67efbd75adc302d24ec3d6ad67ec126f1fcb357d704f735971cc71b569bee2c4\": rpc error: code = NotFound desc = could not find container \"67efbd75adc302d24ec3d6ad67ec126f1fcb357d704f735971cc71b569bee2c4\": container with ID starting with 67efbd75adc302d24ec3d6ad67ec126f1fcb357d704f735971cc71b569bee2c4 not found: ID does not exist" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.582441 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c"] Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.585725 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bbbd659c4-56s2c"] Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.610420 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-client-ca\") pod \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.610474 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-config\") pod \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.610550 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-proxy-ca-bundles\") pod \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.610577 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-serving-cert\") pod \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.610623 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57bjr\" (UniqueName: \"kubernetes.io/projected/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-kube-api-access-57bjr\") pod \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\" (UID: \"84a117a5-7bb6-499f-8b7f-c9f007b86c1d\") " Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.610894 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6vpn\" (UniqueName: \"kubernetes.io/projected/76b2bf27-1695-4652-8469-1b3e34652b5a-kube-api-access-b6vpn\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.610996 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76b2bf27-1695-4652-8469-1b3e34652b5a-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.611088 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76b2bf27-1695-4652-8469-1b3e34652b5a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.611172 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/76b2bf27-1695-4652-8469-1b3e34652b5a-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.611692 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-config" (OuterVolumeSpecName: "config") pod "84a117a5-7bb6-499f-8b7f-c9f007b86c1d" (UID: "84a117a5-7bb6-499f-8b7f-c9f007b86c1d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.611721 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-client-ca" (OuterVolumeSpecName: "client-ca") pod "84a117a5-7bb6-499f-8b7f-c9f007b86c1d" (UID: "84a117a5-7bb6-499f-8b7f-c9f007b86c1d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.611739 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "84a117a5-7bb6-499f-8b7f-c9f007b86c1d" (UID: "84a117a5-7bb6-499f-8b7f-c9f007b86c1d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.613786 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "84a117a5-7bb6-499f-8b7f-c9f007b86c1d" (UID: "84a117a5-7bb6-499f-8b7f-c9f007b86c1d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.613799 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-kube-api-access-57bjr" (OuterVolumeSpecName: "kube-api-access-57bjr") pod "84a117a5-7bb6-499f-8b7f-c9f007b86c1d" (UID: "84a117a5-7bb6-499f-8b7f-c9f007b86c1d"). InnerVolumeSpecName "kube-api-access-57bjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.711976 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57bjr\" (UniqueName: \"kubernetes.io/projected/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-kube-api-access-57bjr\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.712008 4983 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.712018 4983 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.712027 4983 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:11 crc kubenswrapper[4983]: I0318 00:13:11.712035 4983 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84a117a5-7bb6-499f-8b7f-c9f007b86c1d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.548582 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" event={"ID":"84a117a5-7bb6-499f-8b7f-c9f007b86c1d","Type":"ContainerDied","Data":"e05a2c8ce768f95bcd71008cd3e2c91ce02bcbe00ced3e9ff212276c12259417"} Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.548649 4983 scope.go:117] "RemoveContainer" containerID="397d9c5429d1a190839b1b3e27106f17ab442797d9881f2ca9ccfd32584458ec" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.549022 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-577468b5b7-r4qdg" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.582375 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-577468b5b7-r4qdg"] Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.584697 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-577468b5b7-r4qdg"] Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.861556 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76b2bf27-1695-4652-8469-1b3e34652b5a" path="/var/lib/kubelet/pods/76b2bf27-1695-4652-8469-1b3e34652b5a/volumes" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.863008 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84a117a5-7bb6-499f-8b7f-c9f007b86c1d" path="/var/lib/kubelet/pods/84a117a5-7bb6-499f-8b7f-c9f007b86c1d/volumes" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.896771 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd"] Mar 18 00:13:12 crc kubenswrapper[4983]: E0318 00:13:12.897234 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84a117a5-7bb6-499f-8b7f-c9f007b86c1d" containerName="controller-manager" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.897260 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="84a117a5-7bb6-499f-8b7f-c9f007b86c1d" containerName="controller-manager" Mar 18 00:13:12 crc kubenswrapper[4983]: E0318 00:13:12.897297 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76b2bf27-1695-4652-8469-1b3e34652b5a" containerName="route-controller-manager" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.897309 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="76b2bf27-1695-4652-8469-1b3e34652b5a" containerName="route-controller-manager" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.897486 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="84a117a5-7bb6-499f-8b7f-c9f007b86c1d" containerName="controller-manager" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.897508 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="76b2bf27-1695-4652-8469-1b3e34652b5a" containerName="route-controller-manager" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.898149 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.900219 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f"] Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.901103 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.901691 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.902852 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.903077 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.903248 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.903497 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.903718 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.905267 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.905620 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.910029 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.910331 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.910475 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.911273 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.917436 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd"] Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.923412 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f"] Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.932133 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-serving-cert\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.932381 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfc9f849-1672-4b5c-9245-79aaec6111ee-config\") pod \"route-controller-manager-66c9bf54cd-xvwkd\" (UID: \"cfc9f849-1672-4b5c-9245-79aaec6111ee\") " pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.932573 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-config\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.932803 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cfc9f849-1672-4b5c-9245-79aaec6111ee-client-ca\") pod \"route-controller-manager-66c9bf54cd-xvwkd\" (UID: \"cfc9f849-1672-4b5c-9245-79aaec6111ee\") " pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.933013 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfc4g\" (UniqueName: \"kubernetes.io/projected/cfc9f849-1672-4b5c-9245-79aaec6111ee-kube-api-access-qfc4g\") pod \"route-controller-manager-66c9bf54cd-xvwkd\" (UID: \"cfc9f849-1672-4b5c-9245-79aaec6111ee\") " pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.933243 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2tfg\" (UniqueName: \"kubernetes.io/projected/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-kube-api-access-g2tfg\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.933431 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfc9f849-1672-4b5c-9245-79aaec6111ee-serving-cert\") pod \"route-controller-manager-66c9bf54cd-xvwkd\" (UID: \"cfc9f849-1672-4b5c-9245-79aaec6111ee\") " pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.933583 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-proxy-ca-bundles\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.933732 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-client-ca\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:12 crc kubenswrapper[4983]: I0318 00:13:12.933995 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.035200 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfc9f849-1672-4b5c-9245-79aaec6111ee-serving-cert\") pod \"route-controller-manager-66c9bf54cd-xvwkd\" (UID: \"cfc9f849-1672-4b5c-9245-79aaec6111ee\") " pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.035270 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-proxy-ca-bundles\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.035314 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-client-ca\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.035409 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-serving-cert\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.035467 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfc9f849-1672-4b5c-9245-79aaec6111ee-config\") pod \"route-controller-manager-66c9bf54cd-xvwkd\" (UID: \"cfc9f849-1672-4b5c-9245-79aaec6111ee\") " pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.035506 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-config\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.035551 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cfc9f849-1672-4b5c-9245-79aaec6111ee-client-ca\") pod \"route-controller-manager-66c9bf54cd-xvwkd\" (UID: \"cfc9f849-1672-4b5c-9245-79aaec6111ee\") " pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.035586 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfc4g\" (UniqueName: \"kubernetes.io/projected/cfc9f849-1672-4b5c-9245-79aaec6111ee-kube-api-access-qfc4g\") pod \"route-controller-manager-66c9bf54cd-xvwkd\" (UID: \"cfc9f849-1672-4b5c-9245-79aaec6111ee\") " pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.035652 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2tfg\" (UniqueName: \"kubernetes.io/projected/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-kube-api-access-g2tfg\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.036567 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-client-ca\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.036693 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-proxy-ca-bundles\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.037965 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cfc9f849-1672-4b5c-9245-79aaec6111ee-client-ca\") pod \"route-controller-manager-66c9bf54cd-xvwkd\" (UID: \"cfc9f849-1672-4b5c-9245-79aaec6111ee\") " pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.038679 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfc9f849-1672-4b5c-9245-79aaec6111ee-config\") pod \"route-controller-manager-66c9bf54cd-xvwkd\" (UID: \"cfc9f849-1672-4b5c-9245-79aaec6111ee\") " pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.041373 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-config\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.042385 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfc9f849-1672-4b5c-9245-79aaec6111ee-serving-cert\") pod \"route-controller-manager-66c9bf54cd-xvwkd\" (UID: \"cfc9f849-1672-4b5c-9245-79aaec6111ee\") " pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.044127 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-serving-cert\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.058511 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2tfg\" (UniqueName: \"kubernetes.io/projected/2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c-kube-api-access-g2tfg\") pod \"controller-manager-66c7c8f7d7-tlk9f\" (UID: \"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c\") " pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.071548 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfc4g\" (UniqueName: \"kubernetes.io/projected/cfc9f849-1672-4b5c-9245-79aaec6111ee-kube-api-access-qfc4g\") pod \"route-controller-manager-66c9bf54cd-xvwkd\" (UID: \"cfc9f849-1672-4b5c-9245-79aaec6111ee\") " pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.236042 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.257889 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.533757 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd"] Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.557643 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" event={"ID":"cfc9f849-1672-4b5c-9245-79aaec6111ee","Type":"ContainerStarted","Data":"88ba73798305fceba697c49e5698892bddb9b9a8f09a167d032ee3e998ae2c8e"} Mar 18 00:13:13 crc kubenswrapper[4983]: I0318 00:13:13.559024 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f"] Mar 18 00:13:13 crc kubenswrapper[4983]: W0318 00:13:13.563513 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a6dfabb_20d2_4fd3_b2ba_8346bcfdeb4c.slice/crio-31048deb0c0ab26bb660b15256e438eb33c286623b1b37f059430721efbe50b8 WatchSource:0}: Error finding container 31048deb0c0ab26bb660b15256e438eb33c286623b1b37f059430721efbe50b8: Status 404 returned error can't find the container with id 31048deb0c0ab26bb660b15256e438eb33c286623b1b37f059430721efbe50b8 Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.019814 4983 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.020863 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.021014 4983 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.021215 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816" gracePeriod=15 Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.021300 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349" gracePeriod=15 Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.021314 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38" gracePeriod=15 Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.021357 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256" gracePeriod=15 Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.021534 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f" gracePeriod=15 Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.023632 4983 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.025199 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.025309 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.025407 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.025485 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.025563 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.025639 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.025712 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.025791 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.025880 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.025980 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.026075 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.026157 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.026233 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.026309 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.026389 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.026470 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.026550 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.026627 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.026808 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.026891 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.027007 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.027099 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.027182 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.027254 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.027328 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.027412 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.027609 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.027695 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.027871 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.046822 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.047174 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.047313 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.047435 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.047545 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.047659 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.047777 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.047889 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.055352 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.148887 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.148943 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.148987 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.149008 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.149010 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.149021 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.149046 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.149072 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.149089 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.149090 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.149075 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.149100 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.149136 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.149118 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.149118 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.149108 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.353080 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.374378 4983 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.374448 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: W0318 00:13:14.381546 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-08c579adfa5f0f078836e75243c8110201e57a5a59a65bbd54bc17549785d516 WatchSource:0}: Error finding container 08c579adfa5f0f078836e75243c8110201e57a5a59a65bbd54bc17549785d516: Status 404 returned error can't find the container with id 08c579adfa5f0f078836e75243c8110201e57a5a59a65bbd54bc17549785d516 Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.387962 4983 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189dc72606984c79 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:13:14.387172473 +0000 UTC m=+312.284547113,LastTimestamp:2026-03-18 00:13:14.387172473 +0000 UTC m=+312.284547113,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.405645 4983 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.406024 4983 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.406356 4983 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.406679 4983 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.407146 4983 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.407213 4983 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.407590 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="200ms" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.568652 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" event={"ID":"cfc9f849-1672-4b5c-9245-79aaec6111ee","Type":"ContainerStarted","Data":"d68cdaf2a8785faa0d4bbad1e124ce873068ad5745d96af18de1e8c771eda79e"} Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.568851 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.569484 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.569799 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.570120 4983 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.571504 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" event={"ID":"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c","Type":"ContainerStarted","Data":"c3f2233a5d3fac1fa4c74c3ccbd0ddbcf25f730181ab508b644e47b3e435b6c2"} Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.571535 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" event={"ID":"2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c","Type":"ContainerStarted","Data":"31048deb0c0ab26bb660b15256e438eb33c286623b1b37f059430721efbe50b8"} Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.571758 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.572197 4983 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.572562 4983 status_manager.go:851] "Failed to get status for pod" podUID="2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-66c7c8f7d7-tlk9f\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.573171 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.573517 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.575141 4983 generic.go:334] "Generic (PLEG): container finished" podID="551c788d-c0f2-4c95-b728-947250e387ef" containerID="816caf1c5136ad61ac657f0783d3f261737a228d3f657dd7503201c784619ab5" exitCode=0 Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.575221 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"551c788d-c0f2-4c95-b728-947250e387ef","Type":"ContainerDied","Data":"816caf1c5136ad61ac657f0783d3f261737a228d3f657dd7503201c784619ab5"} Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.575705 4983 status_manager.go:851] "Failed to get status for pod" podUID="2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-66c7c8f7d7-tlk9f\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.576101 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.577734 4983 status_manager.go:851] "Failed to get status for pod" podUID="551c788d-c0f2-4c95-b728-947250e387ef" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.578154 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"08c579adfa5f0f078836e75243c8110201e57a5a59a65bbd54bc17549785d516"} Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.578508 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.579009 4983 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.580334 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.580767 4983 status_manager.go:851] "Failed to get status for pod" podUID="2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-66c7c8f7d7-tlk9f\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.581678 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.582101 4983 status_manager.go:851] "Failed to get status for pod" podUID="551c788d-c0f2-4c95-b728-947250e387ef" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.582617 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.583421 4983 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.587813 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.589547 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.591034 4983 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f" exitCode=0 Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.591072 4983 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256" exitCode=0 Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.591102 4983 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38" exitCode=0 Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.591140 4983 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349" exitCode=2 Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.591155 4983 scope.go:117] "RemoveContainer" containerID="b4bda25dd4706fc693337bdf8848ceae5551c18cb558a31865b2444202e84626" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.592664 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.593262 4983 status_manager.go:851] "Failed to get status for pod" podUID="551c788d-c0f2-4c95-b728-947250e387ef" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.594129 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.595387 4983 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.596172 4983 status_manager.go:851] "Failed to get status for pod" podUID="2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-66c7c8f7d7-tlk9f\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: I0318 00:13:14.600155 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:14 crc kubenswrapper[4983]: E0318 00:13:14.608523 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="400ms" Mar 18 00:13:15 crc kubenswrapper[4983]: E0318 00:13:15.009449 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="800ms" Mar 18 00:13:15 crc kubenswrapper[4983]: E0318 00:13:15.396512 4983 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189dc72606984c79 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:13:14.387172473 +0000 UTC m=+312.284547113,LastTimestamp:2026-03-18 00:13:14.387172473 +0000 UTC m=+312.284547113,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:13:15 crc kubenswrapper[4983]: I0318 00:13:15.601895 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"bda654a8732f5608373b25ca99a9841617f1acb6d6e6e8edfe613231e3830d75"} Mar 18 00:13:15 crc kubenswrapper[4983]: I0318 00:13:15.602891 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:15 crc kubenswrapper[4983]: I0318 00:13:15.603455 4983 status_manager.go:851] "Failed to get status for pod" podUID="2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-66c7c8f7d7-tlk9f\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:15 crc kubenswrapper[4983]: I0318 00:13:15.604083 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:15 crc kubenswrapper[4983]: I0318 00:13:15.604609 4983 status_manager.go:851] "Failed to get status for pod" podUID="551c788d-c0f2-4c95-b728-947250e387ef" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:15 crc kubenswrapper[4983]: I0318 00:13:15.606998 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 18 00:13:15 crc kubenswrapper[4983]: E0318 00:13:15.811263 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="1.6s" Mar 18 00:13:15 crc kubenswrapper[4983]: I0318 00:13:15.968814 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 18 00:13:15 crc kubenswrapper[4983]: I0318 00:13:15.969667 4983 status_manager.go:851] "Failed to get status for pod" podUID="2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-66c7c8f7d7-tlk9f\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:15 crc kubenswrapper[4983]: I0318 00:13:15.970424 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:15 crc kubenswrapper[4983]: I0318 00:13:15.971179 4983 status_manager.go:851] "Failed to get status for pod" podUID="551c788d-c0f2-4c95-b728-947250e387ef" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:15 crc kubenswrapper[4983]: I0318 00:13:15.971636 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.075428 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/551c788d-c0f2-4c95-b728-947250e387ef-var-lock\") pod \"551c788d-c0f2-4c95-b728-947250e387ef\" (UID: \"551c788d-c0f2-4c95-b728-947250e387ef\") " Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.075632 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/551c788d-c0f2-4c95-b728-947250e387ef-kube-api-access\") pod \"551c788d-c0f2-4c95-b728-947250e387ef\" (UID: \"551c788d-c0f2-4c95-b728-947250e387ef\") " Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.075700 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/551c788d-c0f2-4c95-b728-947250e387ef-var-lock" (OuterVolumeSpecName: "var-lock") pod "551c788d-c0f2-4c95-b728-947250e387ef" (UID: "551c788d-c0f2-4c95-b728-947250e387ef"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.075711 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/551c788d-c0f2-4c95-b728-947250e387ef-kubelet-dir\") pod \"551c788d-c0f2-4c95-b728-947250e387ef\" (UID: \"551c788d-c0f2-4c95-b728-947250e387ef\") " Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.075760 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/551c788d-c0f2-4c95-b728-947250e387ef-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "551c788d-c0f2-4c95-b728-947250e387ef" (UID: "551c788d-c0f2-4c95-b728-947250e387ef"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.076076 4983 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/551c788d-c0f2-4c95-b728-947250e387ef-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.076112 4983 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/551c788d-c0f2-4c95-b728-947250e387ef-var-lock\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.085127 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/551c788d-c0f2-4c95-b728-947250e387ef-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "551c788d-c0f2-4c95-b728-947250e387ef" (UID: "551c788d-c0f2-4c95-b728-947250e387ef"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.180656 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/551c788d-c0f2-4c95-b728-947250e387ef-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.379763 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.381130 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.381676 4983 status_manager.go:851] "Failed to get status for pod" podUID="2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-66c7c8f7d7-tlk9f\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.382115 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.382784 4983 status_manager.go:851] "Failed to get status for pod" podUID="551c788d-c0f2-4c95-b728-947250e387ef" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.383204 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.383491 4983 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.484058 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.484716 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.484251 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.484832 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.485193 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.485303 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.485957 4983 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.486085 4983 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.486189 4983 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.618363 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"551c788d-c0f2-4c95-b728-947250e387ef","Type":"ContainerDied","Data":"52e4de99e5c15fa8cde703859bdfa8f0c6064f5584fee52294babadf5db9a4a9"} Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.619023 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52e4de99e5c15fa8cde703859bdfa8f0c6064f5584fee52294babadf5db9a4a9" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.618383 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.623990 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.625261 4983 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816" exitCode=0 Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.626080 4983 scope.go:117] "RemoveContainer" containerID="f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.626243 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.640251 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.640739 4983 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.641076 4983 status_manager.go:851] "Failed to get status for pod" podUID="2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-66c7c8f7d7-tlk9f\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.641374 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.641702 4983 status_manager.go:851] "Failed to get status for pod" podUID="551c788d-c0f2-4c95-b728-947250e387ef" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.642143 4983 status_manager.go:851] "Failed to get status for pod" podUID="2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-66c7c8f7d7-tlk9f\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.642367 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.642389 4983 scope.go:117] "RemoveContainer" containerID="3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.642625 4983 status_manager.go:851] "Failed to get status for pod" podUID="551c788d-c0f2-4c95-b728-947250e387ef" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.643385 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.643596 4983 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.662430 4983 scope.go:117] "RemoveContainer" containerID="59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.679603 4983 scope.go:117] "RemoveContainer" containerID="dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.693620 4983 scope.go:117] "RemoveContainer" containerID="29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.710995 4983 scope.go:117] "RemoveContainer" containerID="994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.737898 4983 scope.go:117] "RemoveContainer" containerID="f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f" Mar 18 00:13:16 crc kubenswrapper[4983]: E0318 00:13:16.739710 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\": container with ID starting with f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f not found: ID does not exist" containerID="f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.739756 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f"} err="failed to get container status \"f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\": rpc error: code = NotFound desc = could not find container \"f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f\": container with ID starting with f4e469a30c874c1c18171cd03fdb7c51d758ca429433c24862d641474844eb2f not found: ID does not exist" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.739785 4983 scope.go:117] "RemoveContainer" containerID="3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256" Mar 18 00:13:16 crc kubenswrapper[4983]: E0318 00:13:16.740433 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\": container with ID starting with 3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256 not found: ID does not exist" containerID="3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.740489 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256"} err="failed to get container status \"3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\": rpc error: code = NotFound desc = could not find container \"3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256\": container with ID starting with 3db11481c2ba0864a265b69019513a409827777b93a706ba41f7a13d36302256 not found: ID does not exist" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.740531 4983 scope.go:117] "RemoveContainer" containerID="59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38" Mar 18 00:13:16 crc kubenswrapper[4983]: E0318 00:13:16.740770 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\": container with ID starting with 59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38 not found: ID does not exist" containerID="59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.740802 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38"} err="failed to get container status \"59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\": rpc error: code = NotFound desc = could not find container \"59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38\": container with ID starting with 59539636a9fe9d2378089c8c34948a82223b974336b31244ab45d01b02842a38 not found: ID does not exist" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.740823 4983 scope.go:117] "RemoveContainer" containerID="dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349" Mar 18 00:13:16 crc kubenswrapper[4983]: E0318 00:13:16.741857 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\": container with ID starting with dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349 not found: ID does not exist" containerID="dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.741883 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349"} err="failed to get container status \"dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\": rpc error: code = NotFound desc = could not find container \"dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349\": container with ID starting with dfd2518497741e54c55e066eb909294f6546d447e9c748e4877f64d0fb76d349 not found: ID does not exist" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.741900 4983 scope.go:117] "RemoveContainer" containerID="29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816" Mar 18 00:13:16 crc kubenswrapper[4983]: E0318 00:13:16.742624 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\": container with ID starting with 29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816 not found: ID does not exist" containerID="29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.742752 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816"} err="failed to get container status \"29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\": rpc error: code = NotFound desc = could not find container \"29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816\": container with ID starting with 29d4a9b0d465e5a0b15c586984fc57e231b289ab75456e48e14d0140f6b34816 not found: ID does not exist" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.742781 4983 scope.go:117] "RemoveContainer" containerID="994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716" Mar 18 00:13:16 crc kubenswrapper[4983]: E0318 00:13:16.743118 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\": container with ID starting with 994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716 not found: ID does not exist" containerID="994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.743168 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716"} err="failed to get container status \"994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\": rpc error: code = NotFound desc = could not find container \"994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716\": container with ID starting with 994c4d856f84d4135e7474ba4a39ac319afa7471f9c1b4e58fad5d620cd35716 not found: ID does not exist" Mar 18 00:13:16 crc kubenswrapper[4983]: I0318 00:13:16.857825 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 18 00:13:17 crc kubenswrapper[4983]: E0318 00:13:17.412250 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="3.2s" Mar 18 00:13:20 crc kubenswrapper[4983]: E0318 00:13:20.613401 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="6.4s" Mar 18 00:13:22 crc kubenswrapper[4983]: I0318 00:13:22.850190 4983 status_manager.go:851] "Failed to get status for pod" podUID="551c788d-c0f2-4c95-b728-947250e387ef" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:22 crc kubenswrapper[4983]: I0318 00:13:22.851002 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:22 crc kubenswrapper[4983]: I0318 00:13:22.853297 4983 status_manager.go:851] "Failed to get status for pod" podUID="2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-66c7c8f7d7-tlk9f\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:22 crc kubenswrapper[4983]: I0318 00:13:22.853832 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:25 crc kubenswrapper[4983]: E0318 00:13:25.398104 4983 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189dc72606984c79 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 00:13:14.387172473 +0000 UTC m=+312.284547113,LastTimestamp:2026-03-18 00:13:14.387172473 +0000 UTC m=+312.284547113,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 00:13:25 crc kubenswrapper[4983]: I0318 00:13:25.843724 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:25 crc kubenswrapper[4983]: I0318 00:13:25.846163 4983 status_manager.go:851] "Failed to get status for pod" podUID="551c788d-c0f2-4c95-b728-947250e387ef" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:25 crc kubenswrapper[4983]: I0318 00:13:25.846746 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:25 crc kubenswrapper[4983]: I0318 00:13:25.847595 4983 status_manager.go:851] "Failed to get status for pod" podUID="2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-66c7c8f7d7-tlk9f\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:25 crc kubenswrapper[4983]: I0318 00:13:25.848249 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:25 crc kubenswrapper[4983]: I0318 00:13:25.871524 4983 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2f49d12d-9e58-4eae-a24d-b58a49f81e29" Mar 18 00:13:25 crc kubenswrapper[4983]: I0318 00:13:25.871563 4983 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2f49d12d-9e58-4eae-a24d-b58a49f81e29" Mar 18 00:13:25 crc kubenswrapper[4983]: E0318 00:13:25.872004 4983 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:25 crc kubenswrapper[4983]: I0318 00:13:25.872575 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:25 crc kubenswrapper[4983]: W0318 00:13:25.916451 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-b4460c2754b2c02b96cbbb833bc873069df2449a50ece6efcb6d8c9a3a18b43a WatchSource:0}: Error finding container b4460c2754b2c02b96cbbb833bc873069df2449a50ece6efcb6d8c9a3a18b43a: Status 404 returned error can't find the container with id b4460c2754b2c02b96cbbb833bc873069df2449a50ece6efcb6d8c9a3a18b43a Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.702536 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.704019 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.704116 4983 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2" exitCode=1 Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.704232 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2"} Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.705062 4983 scope.go:117] "RemoveContainer" containerID="cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.705613 4983 status_manager.go:851] "Failed to get status for pod" podUID="2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-66c7c8f7d7-tlk9f\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.706119 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.706403 4983 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.706711 4983 status_manager.go:851] "Failed to get status for pod" podUID="551c788d-c0f2-4c95-b728-947250e387ef" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.707037 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.707417 4983 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="88784613f8e141a3317a84becf0cd14502c0b883dac7c3f3933fb33d2d702117" exitCode=0 Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.707468 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"88784613f8e141a3317a84becf0cd14502c0b883dac7c3f3933fb33d2d702117"} Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.707503 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b4460c2754b2c02b96cbbb833bc873069df2449a50ece6efcb6d8c9a3a18b43a"} Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.707813 4983 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2f49d12d-9e58-4eae-a24d-b58a49f81e29" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.707836 4983 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2f49d12d-9e58-4eae-a24d-b58a49f81e29" Mar 18 00:13:26 crc kubenswrapper[4983]: E0318 00:13:26.708160 4983 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.708372 4983 status_manager.go:851] "Failed to get status for pod" podUID="cfc9f849-1672-4b5c-9245-79aaec6111ee" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-66c9bf54cd-xvwkd\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.708992 4983 status_manager.go:851] "Failed to get status for pod" podUID="551c788d-c0f2-4c95-b728-947250e387ef" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.709373 4983 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.709780 4983 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:26 crc kubenswrapper[4983]: I0318 00:13:26.710399 4983 status_manager.go:851] "Failed to get status for pod" podUID="2a6dfabb-20d2-4fd3-b2ba-8346bcfdeb4c" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-66c7c8f7d7-tlk9f\": dial tcp 38.102.83.217:6443: connect: connection refused" Mar 18 00:13:27 crc kubenswrapper[4983]: E0318 00:13:27.014879 4983 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="7s" Mar 18 00:13:27 crc kubenswrapper[4983]: I0318 00:13:27.728922 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 18 00:13:27 crc kubenswrapper[4983]: I0318 00:13:27.731522 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 18 00:13:27 crc kubenswrapper[4983]: I0318 00:13:27.731703 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3dfc650fd544ddb39dcf0ce408cad22b9caa3dae414e1120e6848fceaaecd6f3"} Mar 18 00:13:27 crc kubenswrapper[4983]: I0318 00:13:27.737697 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c512544214aaac93a5e37e4b1064c8bd5e205fe304af8bfbf07d6031fd9a6e53"} Mar 18 00:13:27 crc kubenswrapper[4983]: I0318 00:13:27.737757 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b360176d422d23a3034c3cd2a73df456982234647ac12d308ac1f99d807e09a8"} Mar 18 00:13:27 crc kubenswrapper[4983]: I0318 00:13:27.737781 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f796950453be0e7a5828d5c6866892a68201c45a16499666d627c88805c1a169"} Mar 18 00:13:28 crc kubenswrapper[4983]: I0318 00:13:28.748620 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"95b0f0f2a5e70467bbd082dabcbe4044aefebeb87c5123b887f1fdfe78655043"} Mar 18 00:13:28 crc kubenswrapper[4983]: I0318 00:13:28.749435 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:28 crc kubenswrapper[4983]: I0318 00:13:28.749526 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e206ade17daeed305c406e34ee36c13184d66c5a1af4818935d8d8019b6078dc"} Mar 18 00:13:28 crc kubenswrapper[4983]: I0318 00:13:28.749190 4983 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2f49d12d-9e58-4eae-a24d-b58a49f81e29" Mar 18 00:13:28 crc kubenswrapper[4983]: I0318 00:13:28.749659 4983 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2f49d12d-9e58-4eae-a24d-b58a49f81e29" Mar 18 00:13:30 crc kubenswrapper[4983]: I0318 00:13:30.873110 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:30 crc kubenswrapper[4983]: I0318 00:13:30.873402 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:30 crc kubenswrapper[4983]: I0318 00:13:30.884491 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:32 crc kubenswrapper[4983]: I0318 00:13:32.824355 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:13:32 crc kubenswrapper[4983]: I0318 00:13:32.824464 4983 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 18 00:13:32 crc kubenswrapper[4983]: I0318 00:13:32.825091 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 18 00:13:33 crc kubenswrapper[4983]: I0318 00:13:33.799968 4983 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:33 crc kubenswrapper[4983]: I0318 00:13:33.927908 4983 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="63009007-ce17-4f23-a502-34d9ea77f4fd" Mar 18 00:13:34 crc kubenswrapper[4983]: I0318 00:13:34.271794 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:13:34 crc kubenswrapper[4983]: I0318 00:13:34.790440 4983 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2f49d12d-9e58-4eae-a24d-b58a49f81e29" Mar 18 00:13:34 crc kubenswrapper[4983]: I0318 00:13:34.790481 4983 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2f49d12d-9e58-4eae-a24d-b58a49f81e29" Mar 18 00:13:34 crc kubenswrapper[4983]: I0318 00:13:34.794565 4983 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="63009007-ce17-4f23-a502-34d9ea77f4fd" Mar 18 00:13:34 crc kubenswrapper[4983]: I0318 00:13:34.795661 4983 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://f796950453be0e7a5828d5c6866892a68201c45a16499666d627c88805c1a169" Mar 18 00:13:34 crc kubenswrapper[4983]: I0318 00:13:34.795705 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:35 crc kubenswrapper[4983]: I0318 00:13:35.795818 4983 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2f49d12d-9e58-4eae-a24d-b58a49f81e29" Mar 18 00:13:35 crc kubenswrapper[4983]: I0318 00:13:35.795864 4983 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2f49d12d-9e58-4eae-a24d-b58a49f81e29" Mar 18 00:13:35 crc kubenswrapper[4983]: I0318 00:13:35.799614 4983 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="63009007-ce17-4f23-a502-34d9ea77f4fd" Mar 18 00:13:42 crc kubenswrapper[4983]: I0318 00:13:42.822657 4983 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 18 00:13:42 crc kubenswrapper[4983]: I0318 00:13:42.823339 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 18 00:13:44 crc kubenswrapper[4983]: I0318 00:13:44.635917 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 18 00:13:44 crc kubenswrapper[4983]: I0318 00:13:44.798180 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 18 00:13:45 crc kubenswrapper[4983]: I0318 00:13:45.256211 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 18 00:13:45 crc kubenswrapper[4983]: I0318 00:13:45.371637 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 18 00:13:45 crc kubenswrapper[4983]: I0318 00:13:45.750664 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 18 00:13:45 crc kubenswrapper[4983]: I0318 00:13:45.915603 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 18 00:13:45 crc kubenswrapper[4983]: I0318 00:13:45.951811 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 18 00:13:46 crc kubenswrapper[4983]: I0318 00:13:46.135483 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 18 00:13:46 crc kubenswrapper[4983]: I0318 00:13:46.309454 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 18 00:13:46 crc kubenswrapper[4983]: I0318 00:13:46.468012 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 18 00:13:46 crc kubenswrapper[4983]: I0318 00:13:46.480697 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 18 00:13:46 crc kubenswrapper[4983]: I0318 00:13:46.514110 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 18 00:13:46 crc kubenswrapper[4983]: I0318 00:13:46.740922 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 18 00:13:46 crc kubenswrapper[4983]: I0318 00:13:46.795806 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 18 00:13:46 crc kubenswrapper[4983]: I0318 00:13:46.840512 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 18 00:13:46 crc kubenswrapper[4983]: I0318 00:13:46.909628 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 18 00:13:46 crc kubenswrapper[4983]: I0318 00:13:46.952188 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 18 00:13:46 crc kubenswrapper[4983]: I0318 00:13:46.981574 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 18 00:13:46 crc kubenswrapper[4983]: I0318 00:13:46.988123 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 18 00:13:47 crc kubenswrapper[4983]: I0318 00:13:47.217543 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 18 00:13:47 crc kubenswrapper[4983]: I0318 00:13:47.222260 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 18 00:13:47 crc kubenswrapper[4983]: I0318 00:13:47.366041 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 18 00:13:47 crc kubenswrapper[4983]: I0318 00:13:47.436067 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 18 00:13:47 crc kubenswrapper[4983]: I0318 00:13:47.470101 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 18 00:13:47 crc kubenswrapper[4983]: I0318 00:13:47.590545 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 18 00:13:47 crc kubenswrapper[4983]: I0318 00:13:47.723812 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 18 00:13:47 crc kubenswrapper[4983]: I0318 00:13:47.729808 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 18 00:13:47 crc kubenswrapper[4983]: I0318 00:13:47.930599 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.136447 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.300527 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.349079 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.371322 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.371496 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.393040 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.425251 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.432018 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.505749 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.540129 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.550759 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.624032 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.746084 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.898853 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 18 00:13:48 crc kubenswrapper[4983]: I0318 00:13:48.983036 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 00:13:49 crc kubenswrapper[4983]: I0318 00:13:49.095785 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 18 00:13:49 crc kubenswrapper[4983]: I0318 00:13:49.098274 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 18 00:13:49 crc kubenswrapper[4983]: I0318 00:13:49.162219 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 18 00:13:49 crc kubenswrapper[4983]: I0318 00:13:49.502124 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 18 00:13:49 crc kubenswrapper[4983]: I0318 00:13:49.666106 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 18 00:13:49 crc kubenswrapper[4983]: I0318 00:13:49.735799 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 18 00:13:49 crc kubenswrapper[4983]: I0318 00:13:49.840105 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 18 00:13:49 crc kubenswrapper[4983]: I0318 00:13:49.884985 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 00:13:49 crc kubenswrapper[4983]: I0318 00:13:49.911149 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 18 00:13:49 crc kubenswrapper[4983]: I0318 00:13:49.990650 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.045379 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.072878 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.183969 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.203907 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.299791 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.328767 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.412322 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.441506 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.459650 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.477223 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.492557 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.534068 4983 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.603228 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.632997 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.643241 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.674702 4983 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.768973 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.938604 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 18 00:13:50 crc kubenswrapper[4983]: I0318 00:13:50.996071 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.104143 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.223865 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.374962 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.394731 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.397246 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.435563 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.465373 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.672665 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.705696 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.707974 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.746583 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.821847 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.849066 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.884496 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.897359 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.925565 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.928756 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 18 00:13:51 crc kubenswrapper[4983]: I0318 00:13:51.983781 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.067501 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.147808 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.199097 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.404462 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.422841 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.463714 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.540160 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.543616 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.614606 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.624776 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.648194 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.648796 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.649574 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.688976 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.710636 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.806031 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.819890 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.822185 4983 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.822452 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.822677 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.823704 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"3dfc650fd544ddb39dcf0ce408cad22b9caa3dae414e1120e6848fceaaecd6f3"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.824089 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://3dfc650fd544ddb39dcf0ce408cad22b9caa3dae414e1120e6848fceaaecd6f3" gracePeriod=30 Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.856380 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.867986 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.905642 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.906051 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 18 00:13:52 crc kubenswrapper[4983]: I0318 00:13:52.994965 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.008571 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.024340 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.108623 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.216601 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.316899 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.337539 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.432796 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.439828 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.583816 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.607455 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.629765 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.651396 4983 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.657845 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.767687 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.819220 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.824591 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.844847 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.956017 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.986690 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 18 00:13:53 crc kubenswrapper[4983]: I0318 00:13:53.997098 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.077342 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.187798 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.231205 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.275496 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.394356 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.434513 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.524096 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.538456 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.642435 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.681432 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.722908 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.779338 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.798896 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.828291 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.870537 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.882057 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.896883 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 18 00:13:54 crc kubenswrapper[4983]: I0318 00:13:54.940381 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.078606 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.118706 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.129728 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.170850 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.191560 4983 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.225597 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.258005 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.315447 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.317564 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.329173 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.343199 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.407916 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.409648 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.428093 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.437084 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.612995 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.615942 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.643542 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.880886 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.900455 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.939246 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.971352 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 18 00:13:55 crc kubenswrapper[4983]: I0318 00:13:55.975064 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.016433 4983 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.022386 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66c7c8f7d7-tlk9f" podStartSLOduration=46.022365412 podStartE2EDuration="46.022365412s" podCreationTimestamp="2026-03-18 00:13:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:13:33.829053101 +0000 UTC m=+331.726427721" watchObservedRunningTime="2026-03-18 00:13:56.022365412 +0000 UTC m=+353.919740062" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.023181 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=42.023172864 podStartE2EDuration="42.023172864s" podCreationTimestamp="2026-03-18 00:13:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:13:33.909473499 +0000 UTC m=+331.806848149" watchObservedRunningTime="2026-03-18 00:13:56.023172864 +0000 UTC m=+353.920547514" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.024428 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-66c9bf54cd-xvwkd" podStartSLOduration=45.024420379 podStartE2EDuration="45.024420379s" podCreationTimestamp="2026-03-18 00:13:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:13:33.852242416 +0000 UTC m=+331.749617036" watchObservedRunningTime="2026-03-18 00:13:56.024420379 +0000 UTC m=+353.921795029" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.025375 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.025430 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.030009 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.045358 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=23.045339991 podStartE2EDuration="23.045339991s" podCreationTimestamp="2026-03-18 00:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:13:56.043594333 +0000 UTC m=+353.940968973" watchObservedRunningTime="2026-03-18 00:13:56.045339991 +0000 UTC m=+353.942714611" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.100414 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.116469 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.216808 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.268449 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.298510 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.359013 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.381637 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.381993 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.427273 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.464660 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.478962 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.498573 4983 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.498800 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://bda654a8732f5608373b25ca99a9841617f1acb6d6e6e8edfe613231e3830d75" gracePeriod=5 Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.639684 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.834722 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.851434 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.866556 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.870302 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.911441 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.914530 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.931218 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 18 00:13:56 crc kubenswrapper[4983]: I0318 00:13:56.973242 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 18 00:13:57 crc kubenswrapper[4983]: I0318 00:13:57.026293 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 18 00:13:57 crc kubenswrapper[4983]: I0318 00:13:57.056210 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 18 00:13:57 crc kubenswrapper[4983]: I0318 00:13:57.061991 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 18 00:13:57 crc kubenswrapper[4983]: I0318 00:13:57.096086 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 18 00:13:57 crc kubenswrapper[4983]: I0318 00:13:57.149648 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 18 00:13:57 crc kubenswrapper[4983]: I0318 00:13:57.155024 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 18 00:13:57 crc kubenswrapper[4983]: I0318 00:13:57.220163 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 18 00:13:57 crc kubenswrapper[4983]: I0318 00:13:57.297302 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 18 00:13:57 crc kubenswrapper[4983]: I0318 00:13:57.404153 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 18 00:13:57 crc kubenswrapper[4983]: I0318 00:13:57.692598 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 18 00:13:57 crc kubenswrapper[4983]: I0318 00:13:57.732664 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 18 00:13:57 crc kubenswrapper[4983]: I0318 00:13:57.790820 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 18 00:13:57 crc kubenswrapper[4983]: I0318 00:13:57.893397 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.101377 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.123695 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.226808 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.314248 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.331440 4983 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.389484 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.411448 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.544234 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.574845 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.675053 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.702560 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.713823 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.753222 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 18 00:13:58 crc kubenswrapper[4983]: I0318 00:13:58.857055 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.073717 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.132247 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.182183 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.247564 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.266518 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.299889 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.333319 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.413143 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.544025 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.667563 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.678303 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.757194 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.944060 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.973371 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 18 00:13:59 crc kubenswrapper[4983]: I0318 00:13:59.996833 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 18 00:14:00 crc kubenswrapper[4983]: I0318 00:14:00.028407 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 18 00:14:00 crc kubenswrapper[4983]: I0318 00:14:00.394311 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 18 00:14:00 crc kubenswrapper[4983]: I0318 00:14:00.588133 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 18 00:14:00 crc kubenswrapper[4983]: I0318 00:14:00.662174 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 18 00:14:00 crc kubenswrapper[4983]: I0318 00:14:00.709600 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 18 00:14:00 crc kubenswrapper[4983]: I0318 00:14:00.874477 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 18 00:14:00 crc kubenswrapper[4983]: I0318 00:14:00.942724 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 18 00:14:01 crc kubenswrapper[4983]: I0318 00:14:01.118791 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 18 00:14:01 crc kubenswrapper[4983]: I0318 00:14:01.356574 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 18 00:14:01 crc kubenswrapper[4983]: I0318 00:14:01.448696 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 18 00:14:01 crc kubenswrapper[4983]: I0318 00:14:01.739987 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 18 00:14:01 crc kubenswrapper[4983]: I0318 00:14:01.926698 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 18 00:14:01 crc kubenswrapper[4983]: I0318 00:14:01.988675 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 18 00:14:01 crc kubenswrapper[4983]: I0318 00:14:01.988734 4983 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="bda654a8732f5608373b25ca99a9841617f1acb6d6e6e8edfe613231e3830d75" exitCode=137 Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.013967 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.081892 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.081971 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.136641 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.136677 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.136701 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.136736 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.136768 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.136795 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.136851 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.136891 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.136938 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.137004 4983 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.137027 4983 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.137037 4983 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.137044 4983 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.151998 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.237571 4983 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.995900 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.996278 4983 scope.go:117] "RemoveContainer" containerID="bda654a8732f5608373b25ca99a9841617f1acb6d6e6e8edfe613231e3830d75" Mar 18 00:14:02 crc kubenswrapper[4983]: I0318 00:14:02.996378 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 00:14:03 crc kubenswrapper[4983]: I0318 00:14:03.010543 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 18 00:14:03 crc kubenswrapper[4983]: I0318 00:14:03.014698 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 18 00:14:03 crc kubenswrapper[4983]: I0318 00:14:03.027043 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 18 00:14:03 crc kubenswrapper[4983]: I0318 00:14:03.027577 4983 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Mar 18 00:14:03 crc kubenswrapper[4983]: I0318 00:14:03.040677 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 18 00:14:03 crc kubenswrapper[4983]: I0318 00:14:03.040734 4983 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="cbd59dc9-b998-4540-ab86-acc31d030ed8" Mar 18 00:14:03 crc kubenswrapper[4983]: I0318 00:14:03.049482 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 18 00:14:03 crc kubenswrapper[4983]: I0318 00:14:03.049544 4983 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="cbd59dc9-b998-4540-ab86-acc31d030ed8" Mar 18 00:14:03 crc kubenswrapper[4983]: I0318 00:14:03.397978 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 18 00:14:23 crc kubenswrapper[4983]: I0318 00:14:23.140185 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 18 00:14:23 crc kubenswrapper[4983]: I0318 00:14:23.142893 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 18 00:14:23 crc kubenswrapper[4983]: I0318 00:14:23.143856 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 18 00:14:23 crc kubenswrapper[4983]: I0318 00:14:23.144010 4983 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="3dfc650fd544ddb39dcf0ce408cad22b9caa3dae414e1120e6848fceaaecd6f3" exitCode=137 Mar 18 00:14:23 crc kubenswrapper[4983]: I0318 00:14:23.144065 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"3dfc650fd544ddb39dcf0ce408cad22b9caa3dae414e1120e6848fceaaecd6f3"} Mar 18 00:14:23 crc kubenswrapper[4983]: I0318 00:14:23.144123 4983 scope.go:117] "RemoveContainer" containerID="cad235613bdf7d49ac733bf82e7919ca380a295db53724eac573b074057da6d2" Mar 18 00:14:24 crc kubenswrapper[4983]: I0318 00:14:24.153046 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 18 00:14:24 crc kubenswrapper[4983]: I0318 00:14:24.154860 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 18 00:14:24 crc kubenswrapper[4983]: I0318 00:14:24.154913 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b0bc3100dcb533646ae14596823349b2682809124dc10d6f237198f647ac5169"} Mar 18 00:14:24 crc kubenswrapper[4983]: I0318 00:14:24.272495 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:14:32 crc kubenswrapper[4983]: I0318 00:14:32.822688 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:14:32 crc kubenswrapper[4983]: I0318 00:14:32.833312 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:14:33 crc kubenswrapper[4983]: I0318 00:14:33.240903 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.177218 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563214-v76f9"] Mar 18 00:14:43 crc kubenswrapper[4983]: E0318 00:14:43.177807 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="551c788d-c0f2-4c95-b728-947250e387ef" containerName="installer" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.177818 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="551c788d-c0f2-4c95-b728-947250e387ef" containerName="installer" Mar 18 00:14:43 crc kubenswrapper[4983]: E0318 00:14:43.177828 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.177835 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.177916 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.177937 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="551c788d-c0f2-4c95-b728-947250e387ef" containerName="installer" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.178427 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563214-v76f9" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.180065 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.180359 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.180839 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.182891 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563214-v76f9"] Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.234111 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc9qt\" (UniqueName: \"kubernetes.io/projected/f97d457f-34d7-4013-9fd5-06cb0cea6720-kube-api-access-nc9qt\") pod \"auto-csr-approver-29563214-v76f9\" (UID: \"f97d457f-34d7-4013-9fd5-06cb0cea6720\") " pod="openshift-infra/auto-csr-approver-29563214-v76f9" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.259003 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.259066 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.335168 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc9qt\" (UniqueName: \"kubernetes.io/projected/f97d457f-34d7-4013-9fd5-06cb0cea6720-kube-api-access-nc9qt\") pod \"auto-csr-approver-29563214-v76f9\" (UID: \"f97d457f-34d7-4013-9fd5-06cb0cea6720\") " pod="openshift-infra/auto-csr-approver-29563214-v76f9" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.361818 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc9qt\" (UniqueName: \"kubernetes.io/projected/f97d457f-34d7-4013-9fd5-06cb0cea6720-kube-api-access-nc9qt\") pod \"auto-csr-approver-29563214-v76f9\" (UID: \"f97d457f-34d7-4013-9fd5-06cb0cea6720\") " pod="openshift-infra/auto-csr-approver-29563214-v76f9" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.492353 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563214-v76f9" Mar 18 00:14:43 crc kubenswrapper[4983]: I0318 00:14:43.929958 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563214-v76f9"] Mar 18 00:14:43 crc kubenswrapper[4983]: W0318 00:14:43.937145 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf97d457f_34d7_4013_9fd5_06cb0cea6720.slice/crio-082cb28bbc4219ba75bf2ecc3dc643624aedb7ea7b5c1a1289c19a8b9012ee07 WatchSource:0}: Error finding container 082cb28bbc4219ba75bf2ecc3dc643624aedb7ea7b5c1a1289c19a8b9012ee07: Status 404 returned error can't find the container with id 082cb28bbc4219ba75bf2ecc3dc643624aedb7ea7b5c1a1289c19a8b9012ee07 Mar 18 00:14:44 crc kubenswrapper[4983]: I0318 00:14:44.308872 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563214-v76f9" event={"ID":"f97d457f-34d7-4013-9fd5-06cb0cea6720","Type":"ContainerStarted","Data":"082cb28bbc4219ba75bf2ecc3dc643624aedb7ea7b5c1a1289c19a8b9012ee07"} Mar 18 00:14:45 crc kubenswrapper[4983]: I0318 00:14:45.317663 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563214-v76f9" event={"ID":"f97d457f-34d7-4013-9fd5-06cb0cea6720","Type":"ContainerStarted","Data":"f600cfda114e5a6574eff91b339af10fdbaf47d0b2edb09e0d6f962bd08dbc36"} Mar 18 00:14:45 crc kubenswrapper[4983]: I0318 00:14:45.334005 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29563214-v76f9" podStartSLOduration=1.4290472379999999 podStartE2EDuration="2.333983331s" podCreationTimestamp="2026-03-18 00:14:43 +0000 UTC" firstStartedPulling="2026-03-18 00:14:43.939959193 +0000 UTC m=+401.837333823" lastFinishedPulling="2026-03-18 00:14:44.844895286 +0000 UTC m=+402.742269916" observedRunningTime="2026-03-18 00:14:45.328376824 +0000 UTC m=+403.225751454" watchObservedRunningTime="2026-03-18 00:14:45.333983331 +0000 UTC m=+403.231357961" Mar 18 00:14:46 crc kubenswrapper[4983]: I0318 00:14:46.325278 4983 generic.go:334] "Generic (PLEG): container finished" podID="f97d457f-34d7-4013-9fd5-06cb0cea6720" containerID="f600cfda114e5a6574eff91b339af10fdbaf47d0b2edb09e0d6f962bd08dbc36" exitCode=0 Mar 18 00:14:46 crc kubenswrapper[4983]: I0318 00:14:46.325341 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563214-v76f9" event={"ID":"f97d457f-34d7-4013-9fd5-06cb0cea6720","Type":"ContainerDied","Data":"f600cfda114e5a6574eff91b339af10fdbaf47d0b2edb09e0d6f962bd08dbc36"} Mar 18 00:14:47 crc kubenswrapper[4983]: E0318 00:14:47.527005 4983 kubelet_node_status.go:756] "Failed to set some node status fields" err="failed to validate nodeIP: route ip+net: no such network interface" node="crc" Mar 18 00:14:47 crc kubenswrapper[4983]: I0318 00:14:47.731244 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563214-v76f9" Mar 18 00:14:47 crc kubenswrapper[4983]: I0318 00:14:47.809344 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc9qt\" (UniqueName: \"kubernetes.io/projected/f97d457f-34d7-4013-9fd5-06cb0cea6720-kube-api-access-nc9qt\") pod \"f97d457f-34d7-4013-9fd5-06cb0cea6720\" (UID: \"f97d457f-34d7-4013-9fd5-06cb0cea6720\") " Mar 18 00:14:47 crc kubenswrapper[4983]: I0318 00:14:47.815063 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f97d457f-34d7-4013-9fd5-06cb0cea6720-kube-api-access-nc9qt" (OuterVolumeSpecName: "kube-api-access-nc9qt") pod "f97d457f-34d7-4013-9fd5-06cb0cea6720" (UID: "f97d457f-34d7-4013-9fd5-06cb0cea6720"). InnerVolumeSpecName "kube-api-access-nc9qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:14:47 crc kubenswrapper[4983]: I0318 00:14:47.910699 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc9qt\" (UniqueName: \"kubernetes.io/projected/f97d457f-34d7-4013-9fd5-06cb0cea6720-kube-api-access-nc9qt\") on node \"crc\" DevicePath \"\"" Mar 18 00:14:48 crc kubenswrapper[4983]: I0318 00:14:48.341335 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563214-v76f9" event={"ID":"f97d457f-34d7-4013-9fd5-06cb0cea6720","Type":"ContainerDied","Data":"082cb28bbc4219ba75bf2ecc3dc643624aedb7ea7b5c1a1289c19a8b9012ee07"} Mar 18 00:14:48 crc kubenswrapper[4983]: I0318 00:14:48.341391 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="082cb28bbc4219ba75bf2ecc3dc643624aedb7ea7b5c1a1289c19a8b9012ee07" Mar 18 00:14:48 crc kubenswrapper[4983]: I0318 00:14:48.341707 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563214-v76f9" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.152833 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h"] Mar 18 00:15:00 crc kubenswrapper[4983]: E0318 00:15:00.153893 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f97d457f-34d7-4013-9fd5-06cb0cea6720" containerName="oc" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.153922 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f97d457f-34d7-4013-9fd5-06cb0cea6720" containerName="oc" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.154176 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f97d457f-34d7-4013-9fd5-06cb0cea6720" containerName="oc" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.154753 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.158325 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.158530 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.167815 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h"] Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.278118 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mhrr\" (UniqueName: \"kubernetes.io/projected/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-kube-api-access-7mhrr\") pod \"collect-profiles-29563215-vwk4h\" (UID: \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.278423 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-config-volume\") pod \"collect-profiles-29563215-vwk4h\" (UID: \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.278571 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-secret-volume\") pod \"collect-profiles-29563215-vwk4h\" (UID: \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.379949 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mhrr\" (UniqueName: \"kubernetes.io/projected/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-kube-api-access-7mhrr\") pod \"collect-profiles-29563215-vwk4h\" (UID: \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.379996 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-config-volume\") pod \"collect-profiles-29563215-vwk4h\" (UID: \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.380057 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-secret-volume\") pod \"collect-profiles-29563215-vwk4h\" (UID: \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.381186 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-config-volume\") pod \"collect-profiles-29563215-vwk4h\" (UID: \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.391242 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-secret-volume\") pod \"collect-profiles-29563215-vwk4h\" (UID: \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.402837 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mhrr\" (UniqueName: \"kubernetes.io/projected/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-kube-api-access-7mhrr\") pod \"collect-profiles-29563215-vwk4h\" (UID: \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" Mar 18 00:15:00 crc kubenswrapper[4983]: I0318 00:15:00.509389 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" Mar 18 00:15:01 crc kubenswrapper[4983]: I0318 00:15:01.010837 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h"] Mar 18 00:15:01 crc kubenswrapper[4983]: W0318 00:15:01.013961 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bd81c43_aea9_4717_b2da_a0fdb8b61e7a.slice/crio-44a007280f029519f476418a48ad377b688f585cb1f64d658d9c9124ef894a8b WatchSource:0}: Error finding container 44a007280f029519f476418a48ad377b688f585cb1f64d658d9c9124ef894a8b: Status 404 returned error can't find the container with id 44a007280f029519f476418a48ad377b688f585cb1f64d658d9c9124ef894a8b Mar 18 00:15:01 crc kubenswrapper[4983]: I0318 00:15:01.426650 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" event={"ID":"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a","Type":"ContainerStarted","Data":"8859f588b57e19b03a1e078f873fd44760b3e6746418e688269a9ca7387eda79"} Mar 18 00:15:01 crc kubenswrapper[4983]: I0318 00:15:01.426706 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" event={"ID":"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a","Type":"ContainerStarted","Data":"44a007280f029519f476418a48ad377b688f585cb1f64d658d9c9124ef894a8b"} Mar 18 00:15:01 crc kubenswrapper[4983]: I0318 00:15:01.441383 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" podStartSLOduration=1.441366372 podStartE2EDuration="1.441366372s" podCreationTimestamp="2026-03-18 00:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:15:01.440614111 +0000 UTC m=+419.337988741" watchObservedRunningTime="2026-03-18 00:15:01.441366372 +0000 UTC m=+419.338740982" Mar 18 00:15:02 crc kubenswrapper[4983]: I0318 00:15:02.437797 4983 generic.go:334] "Generic (PLEG): container finished" podID="8bd81c43-aea9-4717-b2da-a0fdb8b61e7a" containerID="8859f588b57e19b03a1e078f873fd44760b3e6746418e688269a9ca7387eda79" exitCode=0 Mar 18 00:15:02 crc kubenswrapper[4983]: I0318 00:15:02.437870 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" event={"ID":"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a","Type":"ContainerDied","Data":"8859f588b57e19b03a1e078f873fd44760b3e6746418e688269a9ca7387eda79"} Mar 18 00:15:03 crc kubenswrapper[4983]: I0318 00:15:03.820015 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" Mar 18 00:15:03 crc kubenswrapper[4983]: I0318 00:15:03.924085 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-secret-volume\") pod \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\" (UID: \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\") " Mar 18 00:15:03 crc kubenswrapper[4983]: I0318 00:15:03.924211 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-config-volume\") pod \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\" (UID: \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\") " Mar 18 00:15:03 crc kubenswrapper[4983]: I0318 00:15:03.924313 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mhrr\" (UniqueName: \"kubernetes.io/projected/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-kube-api-access-7mhrr\") pod \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\" (UID: \"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a\") " Mar 18 00:15:03 crc kubenswrapper[4983]: I0318 00:15:03.925443 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-config-volume" (OuterVolumeSpecName: "config-volume") pod "8bd81c43-aea9-4717-b2da-a0fdb8b61e7a" (UID: "8bd81c43-aea9-4717-b2da-a0fdb8b61e7a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:15:03 crc kubenswrapper[4983]: I0318 00:15:03.930583 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-kube-api-access-7mhrr" (OuterVolumeSpecName: "kube-api-access-7mhrr") pod "8bd81c43-aea9-4717-b2da-a0fdb8b61e7a" (UID: "8bd81c43-aea9-4717-b2da-a0fdb8b61e7a"). InnerVolumeSpecName "kube-api-access-7mhrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:15:03 crc kubenswrapper[4983]: I0318 00:15:03.931420 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8bd81c43-aea9-4717-b2da-a0fdb8b61e7a" (UID: "8bd81c43-aea9-4717-b2da-a0fdb8b61e7a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:15:04 crc kubenswrapper[4983]: I0318 00:15:04.025603 4983 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:04 crc kubenswrapper[4983]: I0318 00:15:04.025651 4983 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:04 crc kubenswrapper[4983]: I0318 00:15:04.025671 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mhrr\" (UniqueName: \"kubernetes.io/projected/8bd81c43-aea9-4717-b2da-a0fdb8b61e7a-kube-api-access-7mhrr\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:04 crc kubenswrapper[4983]: I0318 00:15:04.454806 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" event={"ID":"8bd81c43-aea9-4717-b2da-a0fdb8b61e7a","Type":"ContainerDied","Data":"44a007280f029519f476418a48ad377b688f585cb1f64d658d9c9124ef894a8b"} Mar 18 00:15:04 crc kubenswrapper[4983]: I0318 00:15:04.454862 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44a007280f029519f476418a48ad377b688f585cb1f64d658d9c9124ef894a8b" Mar 18 00:15:04 crc kubenswrapper[4983]: I0318 00:15:04.454958 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563215-vwk4h" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.108164 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8wdbp"] Mar 18 00:15:08 crc kubenswrapper[4983]: E0318 00:15:08.109112 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bd81c43-aea9-4717-b2da-a0fdb8b61e7a" containerName="collect-profiles" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.109148 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bd81c43-aea9-4717-b2da-a0fdb8b61e7a" containerName="collect-profiles" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.109429 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bd81c43-aea9-4717-b2da-a0fdb8b61e7a" containerName="collect-profiles" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.110275 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.118747 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8wdbp"] Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.220204 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-registry-certificates\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.220267 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-registry-tls\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.220337 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.220360 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-bound-sa-token\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.220390 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.220415 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-trusted-ca\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.220526 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.220626 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hp4z\" (UniqueName: \"kubernetes.io/projected/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-kube-api-access-2hp4z\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.240373 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.321953 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.322027 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-bound-sa-token\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.322069 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-trusted-ca\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.322128 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.322209 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hp4z\" (UniqueName: \"kubernetes.io/projected/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-kube-api-access-2hp4z\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.322278 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-registry-certificates\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.322315 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-registry-tls\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.322846 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.325496 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-registry-certificates\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.326540 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-trusted-ca\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.329809 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-registry-tls\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.333021 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.338701 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-bound-sa-token\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.344863 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hp4z\" (UniqueName: \"kubernetes.io/projected/2b5cc9be-c728-4dd3-8f65-02b5ba07c940-kube-api-access-2hp4z\") pod \"image-registry-66df7c8f76-8wdbp\" (UID: \"2b5cc9be-c728-4dd3-8f65-02b5ba07c940\") " pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.446593 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:08 crc kubenswrapper[4983]: I0318 00:15:08.881854 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8wdbp"] Mar 18 00:15:09 crc kubenswrapper[4983]: I0318 00:15:09.487508 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" event={"ID":"2b5cc9be-c728-4dd3-8f65-02b5ba07c940","Type":"ContainerStarted","Data":"478b585896db8ee0f248ab3feb694c85e87b7352230b74022a4ce9aea7b6114f"} Mar 18 00:15:09 crc kubenswrapper[4983]: I0318 00:15:09.487555 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" event={"ID":"2b5cc9be-c728-4dd3-8f65-02b5ba07c940","Type":"ContainerStarted","Data":"ef2c244a362e3a5db0358bdd4d2e50a843c3f3d12db01d0be9d7747236b335d4"} Mar 18 00:15:09 crc kubenswrapper[4983]: I0318 00:15:09.487698 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:09 crc kubenswrapper[4983]: I0318 00:15:09.517340 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" podStartSLOduration=1.517317774 podStartE2EDuration="1.517317774s" podCreationTimestamp="2026-03-18 00:15:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:15:09.512966451 +0000 UTC m=+427.410341111" watchObservedRunningTime="2026-03-18 00:15:09.517317774 +0000 UTC m=+427.414692394" Mar 18 00:15:13 crc kubenswrapper[4983]: I0318 00:15:13.259291 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:15:13 crc kubenswrapper[4983]: I0318 00:15:13.259714 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:15:28 crc kubenswrapper[4983]: I0318 00:15:28.454099 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-8wdbp" Mar 18 00:15:28 crc kubenswrapper[4983]: I0318 00:15:28.530055 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qkkt2"] Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.259527 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wnc92"] Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.260761 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wnc92" podUID="a81a1b18-1976-45f7-97f5-4dcd7f745af3" containerName="registry-server" containerID="cri-o://eda81797bff5f0f42dc4edec65d6455e9985956ab14c064f4e4970b15110ca7f" gracePeriod=30 Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.290576 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8fpkl"] Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.291327 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8fpkl" podUID="fe54f632-0209-4996-979c-4392b8f7d037" containerName="registry-server" containerID="cri-o://2d6786e68a747e5bc51e202186f8f257e5b430d111b14ffabecb02fccc350f76" gracePeriod=30 Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.302691 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rpftn"] Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.303074 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" podUID="71e981d5-d657-46ef-9ec0-4ddad58be6ee" containerName="marketplace-operator" containerID="cri-o://ab02e9513fbca003c2cfae1f26cd924a07971a7c553f02c70c481bfcd9fdc01a" gracePeriod=30 Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.307960 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lqjwd"] Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.308227 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lqjwd" podUID="5099d24e-b4a3-4990-bc90-980e0188ef3e" containerName="registry-server" containerID="cri-o://34168a0cd52b8e82778cf0197d2d2f3908e74aea5d3fc4a95194966c904725fb" gracePeriod=30 Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.311597 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-78tfj"] Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.311936 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-78tfj" podUID="75077192-24a6-4efb-8484-bea2db4a4812" containerName="registry-server" containerID="cri-o://1019c075b0604a6ba220f5a861fffd2e41468e2909027aeb6459348f11fe7439" gracePeriod=30 Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.315831 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2lhsd"] Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.317579 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.321586 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2lhsd"] Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.451917 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlnf9\" (UniqueName: \"kubernetes.io/projected/0fa6f497-69b3-4033-babc-85299bfeae4b-kube-api-access-rlnf9\") pod \"marketplace-operator-79b997595-2lhsd\" (UID: \"0fa6f497-69b3-4033-babc-85299bfeae4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.452219 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0fa6f497-69b3-4033-babc-85299bfeae4b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2lhsd\" (UID: \"0fa6f497-69b3-4033-babc-85299bfeae4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.452372 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0fa6f497-69b3-4033-babc-85299bfeae4b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2lhsd\" (UID: \"0fa6f497-69b3-4033-babc-85299bfeae4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" Mar 18 00:15:37 crc kubenswrapper[4983]: E0318 00:15:37.542582 4983 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75077192_24a6_4efb_8484_bea2db4a4812.slice/crio-1019c075b0604a6ba220f5a861fffd2e41468e2909027aeb6459348f11fe7439.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75077192_24a6_4efb_8484_bea2db4a4812.slice/crio-conmon-1019c075b0604a6ba220f5a861fffd2e41468e2909027aeb6459348f11fe7439.scope\": RecentStats: unable to find data in memory cache]" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.553282 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0fa6f497-69b3-4033-babc-85299bfeae4b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2lhsd\" (UID: \"0fa6f497-69b3-4033-babc-85299bfeae4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.553329 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0fa6f497-69b3-4033-babc-85299bfeae4b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2lhsd\" (UID: \"0fa6f497-69b3-4033-babc-85299bfeae4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.553365 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlnf9\" (UniqueName: \"kubernetes.io/projected/0fa6f497-69b3-4033-babc-85299bfeae4b-kube-api-access-rlnf9\") pod \"marketplace-operator-79b997595-2lhsd\" (UID: \"0fa6f497-69b3-4033-babc-85299bfeae4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.554740 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0fa6f497-69b3-4033-babc-85299bfeae4b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2lhsd\" (UID: \"0fa6f497-69b3-4033-babc-85299bfeae4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.568135 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0fa6f497-69b3-4033-babc-85299bfeae4b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2lhsd\" (UID: \"0fa6f497-69b3-4033-babc-85299bfeae4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.568772 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlnf9\" (UniqueName: \"kubernetes.io/projected/0fa6f497-69b3-4033-babc-85299bfeae4b-kube-api-access-rlnf9\") pod \"marketplace-operator-79b997595-2lhsd\" (UID: \"0fa6f497-69b3-4033-babc-85299bfeae4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.685940 4983 generic.go:334] "Generic (PLEG): container finished" podID="5099d24e-b4a3-4990-bc90-980e0188ef3e" containerID="34168a0cd52b8e82778cf0197d2d2f3908e74aea5d3fc4a95194966c904725fb" exitCode=0 Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.685961 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lqjwd" event={"ID":"5099d24e-b4a3-4990-bc90-980e0188ef3e","Type":"ContainerDied","Data":"34168a0cd52b8e82778cf0197d2d2f3908e74aea5d3fc4a95194966c904725fb"} Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.688369 4983 generic.go:334] "Generic (PLEG): container finished" podID="fe54f632-0209-4996-979c-4392b8f7d037" containerID="2d6786e68a747e5bc51e202186f8f257e5b430d111b14ffabecb02fccc350f76" exitCode=0 Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.688436 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fpkl" event={"ID":"fe54f632-0209-4996-979c-4392b8f7d037","Type":"ContainerDied","Data":"2d6786e68a747e5bc51e202186f8f257e5b430d111b14ffabecb02fccc350f76"} Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.695330 4983 generic.go:334] "Generic (PLEG): container finished" podID="71e981d5-d657-46ef-9ec0-4ddad58be6ee" containerID="ab02e9513fbca003c2cfae1f26cd924a07971a7c553f02c70c481bfcd9fdc01a" exitCode=0 Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.695410 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" event={"ID":"71e981d5-d657-46ef-9ec0-4ddad58be6ee","Type":"ContainerDied","Data":"ab02e9513fbca003c2cfae1f26cd924a07971a7c553f02c70c481bfcd9fdc01a"} Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.698073 4983 generic.go:334] "Generic (PLEG): container finished" podID="75077192-24a6-4efb-8484-bea2db4a4812" containerID="1019c075b0604a6ba220f5a861fffd2e41468e2909027aeb6459348f11fe7439" exitCode=0 Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.698154 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78tfj" event={"ID":"75077192-24a6-4efb-8484-bea2db4a4812","Type":"ContainerDied","Data":"1019c075b0604a6ba220f5a861fffd2e41468e2909027aeb6459348f11fe7439"} Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.699076 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.709245 4983 generic.go:334] "Generic (PLEG): container finished" podID="a81a1b18-1976-45f7-97f5-4dcd7f745af3" containerID="eda81797bff5f0f42dc4edec65d6455e9985956ab14c064f4e4970b15110ca7f" exitCode=0 Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.709284 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnc92" event={"ID":"a81a1b18-1976-45f7-97f5-4dcd7f745af3","Type":"ContainerDied","Data":"eda81797bff5f0f42dc4edec65d6455e9985956ab14c064f4e4970b15110ca7f"} Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.709323 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnc92" event={"ID":"a81a1b18-1976-45f7-97f5-4dcd7f745af3","Type":"ContainerDied","Data":"e5b75985bcc2c379b15b8dc0572e90ff16ffa2dfa05aaa6ee8e21dc54fb68d45"} Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.709333 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5b75985bcc2c379b15b8dc0572e90ff16ffa2dfa05aaa6ee8e21dc54fb68d45" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.709863 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.755835 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.793652 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.794645 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.794803 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.855834 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5099d24e-b4a3-4990-bc90-980e0188ef3e-catalog-content\") pod \"5099d24e-b4a3-4990-bc90-980e0188ef3e\" (UID: \"5099d24e-b4a3-4990-bc90-980e0188ef3e\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.855865 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a81a1b18-1976-45f7-97f5-4dcd7f745af3-utilities\") pod \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\" (UID: \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.855884 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a81a1b18-1976-45f7-97f5-4dcd7f745af3-catalog-content\") pod \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\" (UID: \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.855909 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75077192-24a6-4efb-8484-bea2db4a4812-utilities\") pod \"75077192-24a6-4efb-8484-bea2db4a4812\" (UID: \"75077192-24a6-4efb-8484-bea2db4a4812\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.855942 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe54f632-0209-4996-979c-4392b8f7d037-catalog-content\") pod \"fe54f632-0209-4996-979c-4392b8f7d037\" (UID: \"fe54f632-0209-4996-979c-4392b8f7d037\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.855972 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5znx9\" (UniqueName: \"kubernetes.io/projected/a81a1b18-1976-45f7-97f5-4dcd7f745af3-kube-api-access-5znx9\") pod \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\" (UID: \"a81a1b18-1976-45f7-97f5-4dcd7f745af3\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.855991 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w8dv\" (UniqueName: \"kubernetes.io/projected/fe54f632-0209-4996-979c-4392b8f7d037-kube-api-access-9w8dv\") pod \"fe54f632-0209-4996-979c-4392b8f7d037\" (UID: \"fe54f632-0209-4996-979c-4392b8f7d037\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.856013 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xjtj\" (UniqueName: \"kubernetes.io/projected/75077192-24a6-4efb-8484-bea2db4a4812-kube-api-access-7xjtj\") pod \"75077192-24a6-4efb-8484-bea2db4a4812\" (UID: \"75077192-24a6-4efb-8484-bea2db4a4812\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.856035 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe54f632-0209-4996-979c-4392b8f7d037-utilities\") pod \"fe54f632-0209-4996-979c-4392b8f7d037\" (UID: \"fe54f632-0209-4996-979c-4392b8f7d037\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.856066 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5099d24e-b4a3-4990-bc90-980e0188ef3e-utilities\") pod \"5099d24e-b4a3-4990-bc90-980e0188ef3e\" (UID: \"5099d24e-b4a3-4990-bc90-980e0188ef3e\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.856084 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66hvg\" (UniqueName: \"kubernetes.io/projected/5099d24e-b4a3-4990-bc90-980e0188ef3e-kube-api-access-66hvg\") pod \"5099d24e-b4a3-4990-bc90-980e0188ef3e\" (UID: \"5099d24e-b4a3-4990-bc90-980e0188ef3e\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.856099 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75077192-24a6-4efb-8484-bea2db4a4812-catalog-content\") pod \"75077192-24a6-4efb-8484-bea2db4a4812\" (UID: \"75077192-24a6-4efb-8484-bea2db4a4812\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.856137 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71e981d5-d657-46ef-9ec0-4ddad58be6ee-marketplace-trusted-ca\") pod \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\" (UID: \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.856161 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nclc8\" (UniqueName: \"kubernetes.io/projected/71e981d5-d657-46ef-9ec0-4ddad58be6ee-kube-api-access-nclc8\") pod \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\" (UID: \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.856197 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/71e981d5-d657-46ef-9ec0-4ddad58be6ee-marketplace-operator-metrics\") pod \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\" (UID: \"71e981d5-d657-46ef-9ec0-4ddad58be6ee\") " Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.857308 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a81a1b18-1976-45f7-97f5-4dcd7f745af3-utilities" (OuterVolumeSpecName: "utilities") pod "a81a1b18-1976-45f7-97f5-4dcd7f745af3" (UID: "a81a1b18-1976-45f7-97f5-4dcd7f745af3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.857429 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe54f632-0209-4996-979c-4392b8f7d037-utilities" (OuterVolumeSpecName: "utilities") pod "fe54f632-0209-4996-979c-4392b8f7d037" (UID: "fe54f632-0209-4996-979c-4392b8f7d037"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.858006 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75077192-24a6-4efb-8484-bea2db4a4812-utilities" (OuterVolumeSpecName: "utilities") pod "75077192-24a6-4efb-8484-bea2db4a4812" (UID: "75077192-24a6-4efb-8484-bea2db4a4812"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.859438 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5099d24e-b4a3-4990-bc90-980e0188ef3e-utilities" (OuterVolumeSpecName: "utilities") pod "5099d24e-b4a3-4990-bc90-980e0188ef3e" (UID: "5099d24e-b4a3-4990-bc90-980e0188ef3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.859906 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71e981d5-d657-46ef-9ec0-4ddad58be6ee-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "71e981d5-d657-46ef-9ec0-4ddad58be6ee" (UID: "71e981d5-d657-46ef-9ec0-4ddad58be6ee"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.868985 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5099d24e-b4a3-4990-bc90-980e0188ef3e-kube-api-access-66hvg" (OuterVolumeSpecName: "kube-api-access-66hvg") pod "5099d24e-b4a3-4990-bc90-980e0188ef3e" (UID: "5099d24e-b4a3-4990-bc90-980e0188ef3e"). InnerVolumeSpecName "kube-api-access-66hvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.870380 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71e981d5-d657-46ef-9ec0-4ddad58be6ee-kube-api-access-nclc8" (OuterVolumeSpecName: "kube-api-access-nclc8") pod "71e981d5-d657-46ef-9ec0-4ddad58be6ee" (UID: "71e981d5-d657-46ef-9ec0-4ddad58be6ee"). InnerVolumeSpecName "kube-api-access-nclc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.871450 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe54f632-0209-4996-979c-4392b8f7d037-kube-api-access-9w8dv" (OuterVolumeSpecName: "kube-api-access-9w8dv") pod "fe54f632-0209-4996-979c-4392b8f7d037" (UID: "fe54f632-0209-4996-979c-4392b8f7d037"). InnerVolumeSpecName "kube-api-access-9w8dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.875468 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a81a1b18-1976-45f7-97f5-4dcd7f745af3-kube-api-access-5znx9" (OuterVolumeSpecName: "kube-api-access-5znx9") pod "a81a1b18-1976-45f7-97f5-4dcd7f745af3" (UID: "a81a1b18-1976-45f7-97f5-4dcd7f745af3"). InnerVolumeSpecName "kube-api-access-5znx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.875527 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75077192-24a6-4efb-8484-bea2db4a4812-kube-api-access-7xjtj" (OuterVolumeSpecName: "kube-api-access-7xjtj") pod "75077192-24a6-4efb-8484-bea2db4a4812" (UID: "75077192-24a6-4efb-8484-bea2db4a4812"). InnerVolumeSpecName "kube-api-access-7xjtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.875957 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e981d5-d657-46ef-9ec0-4ddad58be6ee-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "71e981d5-d657-46ef-9ec0-4ddad58be6ee" (UID: "71e981d5-d657-46ef-9ec0-4ddad58be6ee"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.928886 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5099d24e-b4a3-4990-bc90-980e0188ef3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5099d24e-b4a3-4990-bc90-980e0188ef3e" (UID: "5099d24e-b4a3-4990-bc90-980e0188ef3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.945027 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe54f632-0209-4996-979c-4392b8f7d037-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe54f632-0209-4996-979c-4392b8f7d037" (UID: "fe54f632-0209-4996-979c-4392b8f7d037"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.949577 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a81a1b18-1976-45f7-97f5-4dcd7f745af3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a81a1b18-1976-45f7-97f5-4dcd7f745af3" (UID: "a81a1b18-1976-45f7-97f5-4dcd7f745af3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958131 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5099d24e-b4a3-4990-bc90-980e0188ef3e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958174 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a81a1b18-1976-45f7-97f5-4dcd7f745af3-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958192 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a81a1b18-1976-45f7-97f5-4dcd7f745af3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958217 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75077192-24a6-4efb-8484-bea2db4a4812-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958233 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe54f632-0209-4996-979c-4392b8f7d037-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958250 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5znx9\" (UniqueName: \"kubernetes.io/projected/a81a1b18-1976-45f7-97f5-4dcd7f745af3-kube-api-access-5znx9\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958268 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w8dv\" (UniqueName: \"kubernetes.io/projected/fe54f632-0209-4996-979c-4392b8f7d037-kube-api-access-9w8dv\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958284 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xjtj\" (UniqueName: \"kubernetes.io/projected/75077192-24a6-4efb-8484-bea2db4a4812-kube-api-access-7xjtj\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958298 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe54f632-0209-4996-979c-4392b8f7d037-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958312 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5099d24e-b4a3-4990-bc90-980e0188ef3e-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958327 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66hvg\" (UniqueName: \"kubernetes.io/projected/5099d24e-b4a3-4990-bc90-980e0188ef3e-kube-api-access-66hvg\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958342 4983 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71e981d5-d657-46ef-9ec0-4ddad58be6ee-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958357 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nclc8\" (UniqueName: \"kubernetes.io/projected/71e981d5-d657-46ef-9ec0-4ddad58be6ee-kube-api-access-nclc8\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.958372 4983 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/71e981d5-d657-46ef-9ec0-4ddad58be6ee-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.973304 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2lhsd"] Mar 18 00:15:37 crc kubenswrapper[4983]: I0318 00:15:37.991301 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75077192-24a6-4efb-8484-bea2db4a4812-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75077192-24a6-4efb-8484-bea2db4a4812" (UID: "75077192-24a6-4efb-8484-bea2db4a4812"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.059738 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75077192-24a6-4efb-8484-bea2db4a4812-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.721086 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78tfj" event={"ID":"75077192-24a6-4efb-8484-bea2db4a4812","Type":"ContainerDied","Data":"86be496c37a20089eaba15e2afe206f29958e6e4e36a6be2b9d667e5fe025a9c"} Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.721142 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-78tfj" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.721184 4983 scope.go:117] "RemoveContainer" containerID="1019c075b0604a6ba220f5a861fffd2e41468e2909027aeb6459348f11fe7439" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.725053 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lqjwd" event={"ID":"5099d24e-b4a3-4990-bc90-980e0188ef3e","Type":"ContainerDied","Data":"b30a9dda087848793ea13b5099ed5bdbe2deb683aa09faef3825ec8196490093"} Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.725250 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lqjwd" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.729431 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fpkl" event={"ID":"fe54f632-0209-4996-979c-4392b8f7d037","Type":"ContainerDied","Data":"091dd67e5a9c45c4d7cacd0e54c868bd4c82fde5b4629f06ef7156f6373ec434"} Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.729582 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8fpkl" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.732914 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" event={"ID":"0fa6f497-69b3-4033-babc-85299bfeae4b","Type":"ContainerStarted","Data":"1ed016cb4633bf66c569cec7e055a6b87ab06494717762d6d4e59b791f8e5ed9"} Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.732996 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" event={"ID":"0fa6f497-69b3-4033-babc-85299bfeae4b","Type":"ContainerStarted","Data":"b88cac6d6f9b229619eadcc30d23c5613171ff1398c2cf1726ec014f1e211214"} Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.733298 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.742226 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.744077 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wnc92" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.748845 4983 scope.go:117] "RemoveContainer" containerID="f28a8dc1c6c6da99321c8222daf8d5a4fe3e624a5670cc39e3073c6d7e59035a" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.744074 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rpftn" event={"ID":"71e981d5-d657-46ef-9ec0-4ddad58be6ee","Type":"ContainerDied","Data":"35f502854b30ecc086620c65928c61fcbce8a272d8f90335f219d16f976ab38c"} Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.757358 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.781850 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2lhsd" podStartSLOduration=1.781831548 podStartE2EDuration="1.781831548s" podCreationTimestamp="2026-03-18 00:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:15:38.777812383 +0000 UTC m=+456.675187083" watchObservedRunningTime="2026-03-18 00:15:38.781831548 +0000 UTC m=+456.679206168" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.782326 4983 scope.go:117] "RemoveContainer" containerID="0c48bab5388a827e7984b67764d8bb7410e773eda353b791433bb37068deedc5" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.813090 4983 scope.go:117] "RemoveContainer" containerID="34168a0cd52b8e82778cf0197d2d2f3908e74aea5d3fc4a95194966c904725fb" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.838539 4983 scope.go:117] "RemoveContainer" containerID="78e8203f38fb30d66ba733d9b011647d48d43c26289504b858d26c163f1db6ac" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.865562 4983 scope.go:117] "RemoveContainer" containerID="25781ee37f4c893c4ce77bc9b9f63be73b3ade83b0bae39e0f1ae4ad3f020339" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.866023 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-78tfj"] Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.866067 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-78tfj"] Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.866376 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8fpkl"] Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.869027 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8fpkl"] Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.877719 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wnc92"] Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.881701 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wnc92"] Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.894163 4983 scope.go:117] "RemoveContainer" containerID="2d6786e68a747e5bc51e202186f8f257e5b430d111b14ffabecb02fccc350f76" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.898006 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rpftn"] Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.902779 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rpftn"] Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.910279 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lqjwd"] Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.916060 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lqjwd"] Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.921077 4983 scope.go:117] "RemoveContainer" containerID="8c1f9d5a50870abb84e98864f555ff379749fb5600491b30101a9862a989efd9" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.937236 4983 scope.go:117] "RemoveContainer" containerID="498ac344d5498ae2dea1c12999eae6fe8127ea6d3c5628439d9563533fcf9074" Mar 18 00:15:38 crc kubenswrapper[4983]: I0318 00:15:38.949892 4983 scope.go:117] "RemoveContainer" containerID="ab02e9513fbca003c2cfae1f26cd924a07971a7c553f02c70c481bfcd9fdc01a" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437007 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6q45q"] Mar 18 00:15:39 crc kubenswrapper[4983]: E0318 00:15:39.437299 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e981d5-d657-46ef-9ec0-4ddad58be6ee" containerName="marketplace-operator" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437324 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e981d5-d657-46ef-9ec0-4ddad58be6ee" containerName="marketplace-operator" Mar 18 00:15:39 crc kubenswrapper[4983]: E0318 00:15:39.437339 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a81a1b18-1976-45f7-97f5-4dcd7f745af3" containerName="extract-utilities" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437353 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a81a1b18-1976-45f7-97f5-4dcd7f745af3" containerName="extract-utilities" Mar 18 00:15:39 crc kubenswrapper[4983]: E0318 00:15:39.437365 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5099d24e-b4a3-4990-bc90-980e0188ef3e" containerName="registry-server" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437379 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5099d24e-b4a3-4990-bc90-980e0188ef3e" containerName="registry-server" Mar 18 00:15:39 crc kubenswrapper[4983]: E0318 00:15:39.437395 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75077192-24a6-4efb-8484-bea2db4a4812" containerName="registry-server" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437407 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="75077192-24a6-4efb-8484-bea2db4a4812" containerName="registry-server" Mar 18 00:15:39 crc kubenswrapper[4983]: E0318 00:15:39.437428 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5099d24e-b4a3-4990-bc90-980e0188ef3e" containerName="extract-utilities" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437440 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5099d24e-b4a3-4990-bc90-980e0188ef3e" containerName="extract-utilities" Mar 18 00:15:39 crc kubenswrapper[4983]: E0318 00:15:39.437461 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a81a1b18-1976-45f7-97f5-4dcd7f745af3" containerName="registry-server" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437472 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a81a1b18-1976-45f7-97f5-4dcd7f745af3" containerName="registry-server" Mar 18 00:15:39 crc kubenswrapper[4983]: E0318 00:15:39.437491 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5099d24e-b4a3-4990-bc90-980e0188ef3e" containerName="extract-content" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437503 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5099d24e-b4a3-4990-bc90-980e0188ef3e" containerName="extract-content" Mar 18 00:15:39 crc kubenswrapper[4983]: E0318 00:15:39.437519 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe54f632-0209-4996-979c-4392b8f7d037" containerName="extract-utilities" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437531 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe54f632-0209-4996-979c-4392b8f7d037" containerName="extract-utilities" Mar 18 00:15:39 crc kubenswrapper[4983]: E0318 00:15:39.437550 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75077192-24a6-4efb-8484-bea2db4a4812" containerName="extract-utilities" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437561 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="75077192-24a6-4efb-8484-bea2db4a4812" containerName="extract-utilities" Mar 18 00:15:39 crc kubenswrapper[4983]: E0318 00:15:39.437577 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75077192-24a6-4efb-8484-bea2db4a4812" containerName="extract-content" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437588 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="75077192-24a6-4efb-8484-bea2db4a4812" containerName="extract-content" Mar 18 00:15:39 crc kubenswrapper[4983]: E0318 00:15:39.437601 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe54f632-0209-4996-979c-4392b8f7d037" containerName="registry-server" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437613 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe54f632-0209-4996-979c-4392b8f7d037" containerName="registry-server" Mar 18 00:15:39 crc kubenswrapper[4983]: E0318 00:15:39.437632 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe54f632-0209-4996-979c-4392b8f7d037" containerName="extract-content" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437645 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe54f632-0209-4996-979c-4392b8f7d037" containerName="extract-content" Mar 18 00:15:39 crc kubenswrapper[4983]: E0318 00:15:39.437665 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a81a1b18-1976-45f7-97f5-4dcd7f745af3" containerName="extract-content" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437678 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a81a1b18-1976-45f7-97f5-4dcd7f745af3" containerName="extract-content" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437830 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="75077192-24a6-4efb-8484-bea2db4a4812" containerName="registry-server" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437870 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="71e981d5-d657-46ef-9ec0-4ddad58be6ee" containerName="marketplace-operator" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437888 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="a81a1b18-1976-45f7-97f5-4dcd7f745af3" containerName="registry-server" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437907 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe54f632-0209-4996-979c-4392b8f7d037" containerName="registry-server" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.437977 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="5099d24e-b4a3-4990-bc90-980e0188ef3e" containerName="registry-server" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.439158 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.441818 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.448522 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6q45q"] Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.578908 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rpjb\" (UniqueName: \"kubernetes.io/projected/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-kube-api-access-6rpjb\") pod \"redhat-marketplace-6q45q\" (UID: \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\") " pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.579010 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-utilities\") pod \"redhat-marketplace-6q45q\" (UID: \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\") " pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.579049 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-catalog-content\") pod \"redhat-marketplace-6q45q\" (UID: \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\") " pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.636112 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bqvw2"] Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.637616 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.642072 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.652951 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bqvw2"] Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.679783 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9584343e-c8be-4355-9de2-8ecbd1506da5-utilities\") pod \"redhat-operators-bqvw2\" (UID: \"9584343e-c8be-4355-9de2-8ecbd1506da5\") " pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.679843 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rpjb\" (UniqueName: \"kubernetes.io/projected/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-kube-api-access-6rpjb\") pod \"redhat-marketplace-6q45q\" (UID: \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\") " pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.679878 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9584343e-c8be-4355-9de2-8ecbd1506da5-catalog-content\") pod \"redhat-operators-bqvw2\" (UID: \"9584343e-c8be-4355-9de2-8ecbd1506da5\") " pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.679907 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-utilities\") pod \"redhat-marketplace-6q45q\" (UID: \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\") " pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.679957 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-catalog-content\") pod \"redhat-marketplace-6q45q\" (UID: \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\") " pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.679991 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwdsf\" (UniqueName: \"kubernetes.io/projected/9584343e-c8be-4355-9de2-8ecbd1506da5-kube-api-access-fwdsf\") pod \"redhat-operators-bqvw2\" (UID: \"9584343e-c8be-4355-9de2-8ecbd1506da5\") " pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.680583 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-catalog-content\") pod \"redhat-marketplace-6q45q\" (UID: \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\") " pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.680800 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-utilities\") pod \"redhat-marketplace-6q45q\" (UID: \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\") " pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.709631 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rpjb\" (UniqueName: \"kubernetes.io/projected/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-kube-api-access-6rpjb\") pod \"redhat-marketplace-6q45q\" (UID: \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\") " pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.761048 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.785640 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9584343e-c8be-4355-9de2-8ecbd1506da5-utilities\") pod \"redhat-operators-bqvw2\" (UID: \"9584343e-c8be-4355-9de2-8ecbd1506da5\") " pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.785717 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9584343e-c8be-4355-9de2-8ecbd1506da5-catalog-content\") pod \"redhat-operators-bqvw2\" (UID: \"9584343e-c8be-4355-9de2-8ecbd1506da5\") " pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.785789 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwdsf\" (UniqueName: \"kubernetes.io/projected/9584343e-c8be-4355-9de2-8ecbd1506da5-kube-api-access-fwdsf\") pod \"redhat-operators-bqvw2\" (UID: \"9584343e-c8be-4355-9de2-8ecbd1506da5\") " pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.786428 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9584343e-c8be-4355-9de2-8ecbd1506da5-catalog-content\") pod \"redhat-operators-bqvw2\" (UID: \"9584343e-c8be-4355-9de2-8ecbd1506da5\") " pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.786584 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9584343e-c8be-4355-9de2-8ecbd1506da5-utilities\") pod \"redhat-operators-bqvw2\" (UID: \"9584343e-c8be-4355-9de2-8ecbd1506da5\") " pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.818629 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwdsf\" (UniqueName: \"kubernetes.io/projected/9584343e-c8be-4355-9de2-8ecbd1506da5-kube-api-access-fwdsf\") pod \"redhat-operators-bqvw2\" (UID: \"9584343e-c8be-4355-9de2-8ecbd1506da5\") " pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.954803 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:15:39 crc kubenswrapper[4983]: I0318 00:15:39.959978 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6q45q"] Mar 18 00:15:40 crc kubenswrapper[4983]: I0318 00:15:40.380806 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bqvw2"] Mar 18 00:15:40 crc kubenswrapper[4983]: W0318 00:15:40.385385 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9584343e_c8be_4355_9de2_8ecbd1506da5.slice/crio-68e72973c97efe0b5d392ba128370eea1eba0090a0e68270786dbc0d5aafd404 WatchSource:0}: Error finding container 68e72973c97efe0b5d392ba128370eea1eba0090a0e68270786dbc0d5aafd404: Status 404 returned error can't find the container with id 68e72973c97efe0b5d392ba128370eea1eba0090a0e68270786dbc0d5aafd404 Mar 18 00:15:40 crc kubenswrapper[4983]: I0318 00:15:40.762580 4983 generic.go:334] "Generic (PLEG): container finished" podID="55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" containerID="7b474c7667f4e39bceb7398e335a681e7f66a995c0364d8043ae0d34d174c346" exitCode=0 Mar 18 00:15:40 crc kubenswrapper[4983]: I0318 00:15:40.762630 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6q45q" event={"ID":"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb","Type":"ContainerDied","Data":"7b474c7667f4e39bceb7398e335a681e7f66a995c0364d8043ae0d34d174c346"} Mar 18 00:15:40 crc kubenswrapper[4983]: I0318 00:15:40.762703 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6q45q" event={"ID":"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb","Type":"ContainerStarted","Data":"e6dfd2ffdbb2e56d91a1683388fd4f452da2fa0c1a7d969762a1028d64a13eae"} Mar 18 00:15:40 crc kubenswrapper[4983]: I0318 00:15:40.765698 4983 generic.go:334] "Generic (PLEG): container finished" podID="9584343e-c8be-4355-9de2-8ecbd1506da5" containerID="9752bc3a78296ed10a9fab09724c75ef768f637eb9d4ba3840639abee7597a4b" exitCode=0 Mar 18 00:15:40 crc kubenswrapper[4983]: I0318 00:15:40.765820 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqvw2" event={"ID":"9584343e-c8be-4355-9de2-8ecbd1506da5","Type":"ContainerDied","Data":"9752bc3a78296ed10a9fab09724c75ef768f637eb9d4ba3840639abee7597a4b"} Mar 18 00:15:40 crc kubenswrapper[4983]: I0318 00:15:40.765854 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqvw2" event={"ID":"9584343e-c8be-4355-9de2-8ecbd1506da5","Type":"ContainerStarted","Data":"68e72973c97efe0b5d392ba128370eea1eba0090a0e68270786dbc0d5aafd404"} Mar 18 00:15:40 crc kubenswrapper[4983]: I0318 00:15:40.856545 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5099d24e-b4a3-4990-bc90-980e0188ef3e" path="/var/lib/kubelet/pods/5099d24e-b4a3-4990-bc90-980e0188ef3e/volumes" Mar 18 00:15:40 crc kubenswrapper[4983]: I0318 00:15:40.857211 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71e981d5-d657-46ef-9ec0-4ddad58be6ee" path="/var/lib/kubelet/pods/71e981d5-d657-46ef-9ec0-4ddad58be6ee/volumes" Mar 18 00:15:40 crc kubenswrapper[4983]: I0318 00:15:40.857622 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75077192-24a6-4efb-8484-bea2db4a4812" path="/var/lib/kubelet/pods/75077192-24a6-4efb-8484-bea2db4a4812/volumes" Mar 18 00:15:40 crc kubenswrapper[4983]: I0318 00:15:40.858223 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a81a1b18-1976-45f7-97f5-4dcd7f745af3" path="/var/lib/kubelet/pods/a81a1b18-1976-45f7-97f5-4dcd7f745af3/volumes" Mar 18 00:15:40 crc kubenswrapper[4983]: I0318 00:15:40.858757 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe54f632-0209-4996-979c-4392b8f7d037" path="/var/lib/kubelet/pods/fe54f632-0209-4996-979c-4392b8f7d037/volumes" Mar 18 00:15:41 crc kubenswrapper[4983]: I0318 00:15:41.775049 4983 generic.go:334] "Generic (PLEG): container finished" podID="55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" containerID="d9643f6c4875b52c89c8525ecb86ec72bc696d7ea7167509255edcf1120effc2" exitCode=0 Mar 18 00:15:41 crc kubenswrapper[4983]: I0318 00:15:41.775145 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6q45q" event={"ID":"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb","Type":"ContainerDied","Data":"d9643f6c4875b52c89c8525ecb86ec72bc696d7ea7167509255edcf1120effc2"} Mar 18 00:15:41 crc kubenswrapper[4983]: I0318 00:15:41.779111 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqvw2" event={"ID":"9584343e-c8be-4355-9de2-8ecbd1506da5","Type":"ContainerStarted","Data":"ab036da5a088e4b76a3733da8fe94fdcf29413989c3a9fdd1440ac18ceaef931"} Mar 18 00:15:41 crc kubenswrapper[4983]: I0318 00:15:41.834776 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7scqt"] Mar 18 00:15:41 crc kubenswrapper[4983]: I0318 00:15:41.835968 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:41 crc kubenswrapper[4983]: I0318 00:15:41.837703 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 18 00:15:41 crc kubenswrapper[4983]: I0318 00:15:41.841898 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7scqt"] Mar 18 00:15:41 crc kubenswrapper[4983]: I0318 00:15:41.913248 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs42c\" (UniqueName: \"kubernetes.io/projected/970f8e08-cc08-449e-974e-9ae2f62146f7-kube-api-access-zs42c\") pod \"certified-operators-7scqt\" (UID: \"970f8e08-cc08-449e-974e-9ae2f62146f7\") " pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:41 crc kubenswrapper[4983]: I0318 00:15:41.913354 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/970f8e08-cc08-449e-974e-9ae2f62146f7-catalog-content\") pod \"certified-operators-7scqt\" (UID: \"970f8e08-cc08-449e-974e-9ae2f62146f7\") " pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:41 crc kubenswrapper[4983]: I0318 00:15:41.913396 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/970f8e08-cc08-449e-974e-9ae2f62146f7-utilities\") pod \"certified-operators-7scqt\" (UID: \"970f8e08-cc08-449e-974e-9ae2f62146f7\") " pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.014821 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs42c\" (UniqueName: \"kubernetes.io/projected/970f8e08-cc08-449e-974e-9ae2f62146f7-kube-api-access-zs42c\") pod \"certified-operators-7scqt\" (UID: \"970f8e08-cc08-449e-974e-9ae2f62146f7\") " pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.014948 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/970f8e08-cc08-449e-974e-9ae2f62146f7-catalog-content\") pod \"certified-operators-7scqt\" (UID: \"970f8e08-cc08-449e-974e-9ae2f62146f7\") " pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.014993 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/970f8e08-cc08-449e-974e-9ae2f62146f7-utilities\") pod \"certified-operators-7scqt\" (UID: \"970f8e08-cc08-449e-974e-9ae2f62146f7\") " pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.015736 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/970f8e08-cc08-449e-974e-9ae2f62146f7-utilities\") pod \"certified-operators-7scqt\" (UID: \"970f8e08-cc08-449e-974e-9ae2f62146f7\") " pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.015876 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/970f8e08-cc08-449e-974e-9ae2f62146f7-catalog-content\") pod \"certified-operators-7scqt\" (UID: \"970f8e08-cc08-449e-974e-9ae2f62146f7\") " pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.040640 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v96xz"] Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.042339 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.046238 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.054046 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs42c\" (UniqueName: \"kubernetes.io/projected/970f8e08-cc08-449e-974e-9ae2f62146f7-kube-api-access-zs42c\") pod \"certified-operators-7scqt\" (UID: \"970f8e08-cc08-449e-974e-9ae2f62146f7\") " pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.057518 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v96xz"] Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.116519 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwn94\" (UniqueName: \"kubernetes.io/projected/9521010b-fae6-4407-8411-585c5a2391dc-kube-api-access-jwn94\") pod \"community-operators-v96xz\" (UID: \"9521010b-fae6-4407-8411-585c5a2391dc\") " pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.116579 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9521010b-fae6-4407-8411-585c5a2391dc-utilities\") pod \"community-operators-v96xz\" (UID: \"9521010b-fae6-4407-8411-585c5a2391dc\") " pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.116624 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9521010b-fae6-4407-8411-585c5a2391dc-catalog-content\") pod \"community-operators-v96xz\" (UID: \"9521010b-fae6-4407-8411-585c5a2391dc\") " pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.151271 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.217765 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9521010b-fae6-4407-8411-585c5a2391dc-utilities\") pod \"community-operators-v96xz\" (UID: \"9521010b-fae6-4407-8411-585c5a2391dc\") " pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.217826 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9521010b-fae6-4407-8411-585c5a2391dc-catalog-content\") pod \"community-operators-v96xz\" (UID: \"9521010b-fae6-4407-8411-585c5a2391dc\") " pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.217879 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwn94\" (UniqueName: \"kubernetes.io/projected/9521010b-fae6-4407-8411-585c5a2391dc-kube-api-access-jwn94\") pod \"community-operators-v96xz\" (UID: \"9521010b-fae6-4407-8411-585c5a2391dc\") " pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.218503 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9521010b-fae6-4407-8411-585c5a2391dc-catalog-content\") pod \"community-operators-v96xz\" (UID: \"9521010b-fae6-4407-8411-585c5a2391dc\") " pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.218868 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9521010b-fae6-4407-8411-585c5a2391dc-utilities\") pod \"community-operators-v96xz\" (UID: \"9521010b-fae6-4407-8411-585c5a2391dc\") " pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.250805 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwn94\" (UniqueName: \"kubernetes.io/projected/9521010b-fae6-4407-8411-585c5a2391dc-kube-api-access-jwn94\") pod \"community-operators-v96xz\" (UID: \"9521010b-fae6-4407-8411-585c5a2391dc\") " pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.378770 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7scqt"] Mar 18 00:15:42 crc kubenswrapper[4983]: W0318 00:15:42.389779 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod970f8e08_cc08_449e_974e_9ae2f62146f7.slice/crio-24627a5f603c6c8e21948d504af779f018745770f407b010e9f435670189577a WatchSource:0}: Error finding container 24627a5f603c6c8e21948d504af779f018745770f407b010e9f435670189577a: Status 404 returned error can't find the container with id 24627a5f603c6c8e21948d504af779f018745770f407b010e9f435670189577a Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.398151 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.663467 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v96xz"] Mar 18 00:15:42 crc kubenswrapper[4983]: W0318 00:15:42.675317 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9521010b_fae6_4407_8411_585c5a2391dc.slice/crio-9c4d87b967bf3585ef69e140a10b8ee944460e3be93e7a3124211bfd21c50e62 WatchSource:0}: Error finding container 9c4d87b967bf3585ef69e140a10b8ee944460e3be93e7a3124211bfd21c50e62: Status 404 returned error can't find the container with id 9c4d87b967bf3585ef69e140a10b8ee944460e3be93e7a3124211bfd21c50e62 Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.787080 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v96xz" event={"ID":"9521010b-fae6-4407-8411-585c5a2391dc","Type":"ContainerStarted","Data":"9c4d87b967bf3585ef69e140a10b8ee944460e3be93e7a3124211bfd21c50e62"} Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.790026 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6q45q" event={"ID":"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb","Type":"ContainerStarted","Data":"d0f900680a081dabf19a562ca3dfce551f367d448b3d71252a6134f65cedb225"} Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.794222 4983 generic.go:334] "Generic (PLEG): container finished" podID="9584343e-c8be-4355-9de2-8ecbd1506da5" containerID="ab036da5a088e4b76a3733da8fe94fdcf29413989c3a9fdd1440ac18ceaef931" exitCode=0 Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.794281 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqvw2" event={"ID":"9584343e-c8be-4355-9de2-8ecbd1506da5","Type":"ContainerDied","Data":"ab036da5a088e4b76a3733da8fe94fdcf29413989c3a9fdd1440ac18ceaef931"} Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.795996 4983 generic.go:334] "Generic (PLEG): container finished" podID="970f8e08-cc08-449e-974e-9ae2f62146f7" containerID="d114696e2b8b65f4481a70e2e6acd88fc6badb366c32ee7bc24c55930138143d" exitCode=0 Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.796021 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7scqt" event={"ID":"970f8e08-cc08-449e-974e-9ae2f62146f7","Type":"ContainerDied","Data":"d114696e2b8b65f4481a70e2e6acd88fc6badb366c32ee7bc24c55930138143d"} Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.796038 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7scqt" event={"ID":"970f8e08-cc08-449e-974e-9ae2f62146f7","Type":"ContainerStarted","Data":"24627a5f603c6c8e21948d504af779f018745770f407b010e9f435670189577a"} Mar 18 00:15:42 crc kubenswrapper[4983]: I0318 00:15:42.825061 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6q45q" podStartSLOduration=2.145243723 podStartE2EDuration="3.824992061s" podCreationTimestamp="2026-03-18 00:15:39 +0000 UTC" firstStartedPulling="2026-03-18 00:15:40.764776573 +0000 UTC m=+458.662151183" lastFinishedPulling="2026-03-18 00:15:42.444524911 +0000 UTC m=+460.341899521" observedRunningTime="2026-03-18 00:15:42.824089854 +0000 UTC m=+460.721464484" watchObservedRunningTime="2026-03-18 00:15:42.824992061 +0000 UTC m=+460.722366701" Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.259249 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.259326 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.259387 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.260248 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e5e3b4eea0e0cda0f8f16612d694e8e56064b538ff4f0543482bc9f225a5195"} pod="openshift-machine-config-operator/machine-config-daemon-jczfs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.260354 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" containerID="cri-o://7e5e3b4eea0e0cda0f8f16612d694e8e56064b538ff4f0543482bc9f225a5195" gracePeriod=600 Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.804882 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqvw2" event={"ID":"9584343e-c8be-4355-9de2-8ecbd1506da5","Type":"ContainerStarted","Data":"e018842523f2f9a22290c898a30dd852285365243bb6848376b205ae5e77f97a"} Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.807668 4983 generic.go:334] "Generic (PLEG): container finished" podID="970f8e08-cc08-449e-974e-9ae2f62146f7" containerID="04e820adaacdda46c3d66d45614c77acd648017eae9fac5e9688bde8e24f0a53" exitCode=0 Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.808137 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7scqt" event={"ID":"970f8e08-cc08-449e-974e-9ae2f62146f7","Type":"ContainerDied","Data":"04e820adaacdda46c3d66d45614c77acd648017eae9fac5e9688bde8e24f0a53"} Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.810006 4983 generic.go:334] "Generic (PLEG): container finished" podID="9521010b-fae6-4407-8411-585c5a2391dc" containerID="0a040e279a8bfb6372808301b5b61d4e71843b1527c1171362b3fd7f0683ff84" exitCode=0 Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.810032 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v96xz" event={"ID":"9521010b-fae6-4407-8411-585c5a2391dc","Type":"ContainerDied","Data":"0a040e279a8bfb6372808301b5b61d4e71843b1527c1171362b3fd7f0683ff84"} Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.816303 4983 generic.go:334] "Generic (PLEG): container finished" podID="482e2afe-48be-4ea2-8bda-52849c901b52" containerID="7e5e3b4eea0e0cda0f8f16612d694e8e56064b538ff4f0543482bc9f225a5195" exitCode=0 Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.816339 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerDied","Data":"7e5e3b4eea0e0cda0f8f16612d694e8e56064b538ff4f0543482bc9f225a5195"} Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.816382 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerStarted","Data":"7d469517a2f6224bd41c05c725e6f6ca91ae6c676622fc45a16b3a323fcf16e9"} Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.816399 4983 scope.go:117] "RemoveContainer" containerID="ee6c526561d7edb990bc03214d2c851816593e5e175965f8227d33e663de2062" Mar 18 00:15:43 crc kubenswrapper[4983]: I0318 00:15:43.829804 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bqvw2" podStartSLOduration=2.333440152 podStartE2EDuration="4.829788559s" podCreationTimestamp="2026-03-18 00:15:39 +0000 UTC" firstStartedPulling="2026-03-18 00:15:40.768695707 +0000 UTC m=+458.666070317" lastFinishedPulling="2026-03-18 00:15:43.265044114 +0000 UTC m=+461.162418724" observedRunningTime="2026-03-18 00:15:43.823813127 +0000 UTC m=+461.721187777" watchObservedRunningTime="2026-03-18 00:15:43.829788559 +0000 UTC m=+461.727163169" Mar 18 00:15:44 crc kubenswrapper[4983]: I0318 00:15:44.839883 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7scqt" event={"ID":"970f8e08-cc08-449e-974e-9ae2f62146f7","Type":"ContainerStarted","Data":"55a90ab7b72d050763961b2eb044ff16abc89d0245837a9da85958bd6e36b0d4"} Mar 18 00:15:44 crc kubenswrapper[4983]: I0318 00:15:44.878524 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7scqt" podStartSLOduration=2.355008907 podStartE2EDuration="3.878492063s" podCreationTimestamp="2026-03-18 00:15:41 +0000 UTC" firstStartedPulling="2026-03-18 00:15:42.797826028 +0000 UTC m=+460.695200638" lastFinishedPulling="2026-03-18 00:15:44.321309164 +0000 UTC m=+462.218683794" observedRunningTime="2026-03-18 00:15:44.861476407 +0000 UTC m=+462.758851037" watchObservedRunningTime="2026-03-18 00:15:44.878492063 +0000 UTC m=+462.775866673" Mar 18 00:15:45 crc kubenswrapper[4983]: I0318 00:15:45.858124 4983 generic.go:334] "Generic (PLEG): container finished" podID="9521010b-fae6-4407-8411-585c5a2391dc" containerID="d5f8bea7c342351d0acb8341b6a4453cbfeefe97e7e5cd914a1ac2877a125f82" exitCode=0 Mar 18 00:15:45 crc kubenswrapper[4983]: I0318 00:15:45.858189 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v96xz" event={"ID":"9521010b-fae6-4407-8411-585c5a2391dc","Type":"ContainerDied","Data":"d5f8bea7c342351d0acb8341b6a4453cbfeefe97e7e5cd914a1ac2877a125f82"} Mar 18 00:15:46 crc kubenswrapper[4983]: I0318 00:15:46.878893 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v96xz" event={"ID":"9521010b-fae6-4407-8411-585c5a2391dc","Type":"ContainerStarted","Data":"a0b32762e1b50b883cac04f96ea68f37e663951aa59f7b65fe10d22a58ad13ed"} Mar 18 00:15:46 crc kubenswrapper[4983]: I0318 00:15:46.909912 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v96xz" podStartSLOduration=2.471650896 podStartE2EDuration="4.909890464s" podCreationTimestamp="2026-03-18 00:15:42 +0000 UTC" firstStartedPulling="2026-03-18 00:15:43.811475601 +0000 UTC m=+461.708850211" lastFinishedPulling="2026-03-18 00:15:46.249715159 +0000 UTC m=+464.147089779" observedRunningTime="2026-03-18 00:15:46.901663663 +0000 UTC m=+464.799038283" watchObservedRunningTime="2026-03-18 00:15:46.909890464 +0000 UTC m=+464.807265084" Mar 18 00:15:49 crc kubenswrapper[4983]: I0318 00:15:49.762279 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:49 crc kubenswrapper[4983]: I0318 00:15:49.762769 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:49 crc kubenswrapper[4983]: I0318 00:15:49.835111 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:49 crc kubenswrapper[4983]: I0318 00:15:49.955098 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:15:49 crc kubenswrapper[4983]: I0318 00:15:49.955340 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:15:49 crc kubenswrapper[4983]: I0318 00:15:49.970117 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:15:51 crc kubenswrapper[4983]: I0318 00:15:51.008566 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bqvw2" podUID="9584343e-c8be-4355-9de2-8ecbd1506da5" containerName="registry-server" probeResult="failure" output=< Mar 18 00:15:51 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Mar 18 00:15:51 crc kubenswrapper[4983]: > Mar 18 00:15:52 crc kubenswrapper[4983]: I0318 00:15:52.151959 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:52 crc kubenswrapper[4983]: I0318 00:15:52.153575 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:52 crc kubenswrapper[4983]: I0318 00:15:52.216434 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:52 crc kubenswrapper[4983]: I0318 00:15:52.399302 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:52 crc kubenswrapper[4983]: I0318 00:15:52.399365 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:52 crc kubenswrapper[4983]: I0318 00:15:52.471197 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:52 crc kubenswrapper[4983]: I0318 00:15:52.990276 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v96xz" Mar 18 00:15:52 crc kubenswrapper[4983]: I0318 00:15:52.991323 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7scqt" Mar 18 00:15:53 crc kubenswrapper[4983]: I0318 00:15:53.596688 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" podUID="297c1426-c499-449c-ab44-e1f898262839" containerName="registry" containerID="cri-o://36ac6c57e9ecbf8bed1f3803e76cd664da2818fb62f8e12aef88b060b41af978" gracePeriod=30 Mar 18 00:15:53 crc kubenswrapper[4983]: I0318 00:15:53.928844 4983 generic.go:334] "Generic (PLEG): container finished" podID="297c1426-c499-449c-ab44-e1f898262839" containerID="36ac6c57e9ecbf8bed1f3803e76cd664da2818fb62f8e12aef88b060b41af978" exitCode=0 Mar 18 00:15:53 crc kubenswrapper[4983]: I0318 00:15:53.928978 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" event={"ID":"297c1426-c499-449c-ab44-e1f898262839","Type":"ContainerDied","Data":"36ac6c57e9ecbf8bed1f3803e76cd664da2818fb62f8e12aef88b060b41af978"} Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.047051 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.206554 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/297c1426-c499-449c-ab44-e1f898262839-registry-certificates\") pod \"297c1426-c499-449c-ab44-e1f898262839\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.206629 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6grh\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-kube-api-access-l6grh\") pod \"297c1426-c499-449c-ab44-e1f898262839\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.206871 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"297c1426-c499-449c-ab44-e1f898262839\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.206973 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/297c1426-c499-449c-ab44-e1f898262839-installation-pull-secrets\") pod \"297c1426-c499-449c-ab44-e1f898262839\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.207113 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/297c1426-c499-449c-ab44-e1f898262839-ca-trust-extracted\") pod \"297c1426-c499-449c-ab44-e1f898262839\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.207166 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/297c1426-c499-449c-ab44-e1f898262839-trusted-ca\") pod \"297c1426-c499-449c-ab44-e1f898262839\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.207216 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-registry-tls\") pod \"297c1426-c499-449c-ab44-e1f898262839\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.207256 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-bound-sa-token\") pod \"297c1426-c499-449c-ab44-e1f898262839\" (UID: \"297c1426-c499-449c-ab44-e1f898262839\") " Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.208656 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/297c1426-c499-449c-ab44-e1f898262839-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "297c1426-c499-449c-ab44-e1f898262839" (UID: "297c1426-c499-449c-ab44-e1f898262839"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.209307 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/297c1426-c499-449c-ab44-e1f898262839-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "297c1426-c499-449c-ab44-e1f898262839" (UID: "297c1426-c499-449c-ab44-e1f898262839"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.214562 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "297c1426-c499-449c-ab44-e1f898262839" (UID: "297c1426-c499-449c-ab44-e1f898262839"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.216434 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-kube-api-access-l6grh" (OuterVolumeSpecName: "kube-api-access-l6grh") pod "297c1426-c499-449c-ab44-e1f898262839" (UID: "297c1426-c499-449c-ab44-e1f898262839"). InnerVolumeSpecName "kube-api-access-l6grh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.216994 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/297c1426-c499-449c-ab44-e1f898262839-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "297c1426-c499-449c-ab44-e1f898262839" (UID: "297c1426-c499-449c-ab44-e1f898262839"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.217218 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "297c1426-c499-449c-ab44-e1f898262839" (UID: "297c1426-c499-449c-ab44-e1f898262839"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.231655 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/297c1426-c499-449c-ab44-e1f898262839-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "297c1426-c499-449c-ab44-e1f898262839" (UID: "297c1426-c499-449c-ab44-e1f898262839"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.278269 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "297c1426-c499-449c-ab44-e1f898262839" (UID: "297c1426-c499-449c-ab44-e1f898262839"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.309849 4983 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/297c1426-c499-449c-ab44-e1f898262839-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.309879 4983 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/297c1426-c499-449c-ab44-e1f898262839-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.309889 4983 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/297c1426-c499-449c-ab44-e1f898262839-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.309898 4983 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.309907 4983 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.309914 4983 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/297c1426-c499-449c-ab44-e1f898262839-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.309922 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6grh\" (UniqueName: \"kubernetes.io/projected/297c1426-c499-449c-ab44-e1f898262839-kube-api-access-l6grh\") on node \"crc\" DevicePath \"\"" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.937511 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.938147 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qkkt2" event={"ID":"297c1426-c499-449c-ab44-e1f898262839","Type":"ContainerDied","Data":"37776685707eb97ab07718f65b6b31581627b8cc12e80ea52a82dafa2914945b"} Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.938180 4983 scope.go:117] "RemoveContainer" containerID="36ac6c57e9ecbf8bed1f3803e76cd664da2818fb62f8e12aef88b060b41af978" Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.962146 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qkkt2"] Mar 18 00:15:54 crc kubenswrapper[4983]: I0318 00:15:54.969735 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qkkt2"] Mar 18 00:15:56 crc kubenswrapper[4983]: I0318 00:15:56.870884 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="297c1426-c499-449c-ab44-e1f898262839" path="/var/lib/kubelet/pods/297c1426-c499-449c-ab44-e1f898262839/volumes" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.022125 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.102857 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bqvw2" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.153654 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563216-jj8fm"] Mar 18 00:16:00 crc kubenswrapper[4983]: E0318 00:16:00.154102 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="297c1426-c499-449c-ab44-e1f898262839" containerName="registry" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.154144 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="297c1426-c499-449c-ab44-e1f898262839" containerName="registry" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.154366 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="297c1426-c499-449c-ab44-e1f898262839" containerName="registry" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.155119 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563216-jj8fm" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.158467 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.159045 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.159411 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.170093 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563216-jj8fm"] Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.183875 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjqwn\" (UniqueName: \"kubernetes.io/projected/a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf-kube-api-access-xjqwn\") pod \"auto-csr-approver-29563216-jj8fm\" (UID: \"a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf\") " pod="openshift-infra/auto-csr-approver-29563216-jj8fm" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.284825 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjqwn\" (UniqueName: \"kubernetes.io/projected/a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf-kube-api-access-xjqwn\") pod \"auto-csr-approver-29563216-jj8fm\" (UID: \"a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf\") " pod="openshift-infra/auto-csr-approver-29563216-jj8fm" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.318194 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjqwn\" (UniqueName: \"kubernetes.io/projected/a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf-kube-api-access-xjqwn\") pod \"auto-csr-approver-29563216-jj8fm\" (UID: \"a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf\") " pod="openshift-infra/auto-csr-approver-29563216-jj8fm" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.472352 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563216-jj8fm" Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.770220 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563216-jj8fm"] Mar 18 00:16:00 crc kubenswrapper[4983]: I0318 00:16:00.992340 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563216-jj8fm" event={"ID":"a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf","Type":"ContainerStarted","Data":"3a165dbce117b658b4e83612be5864cdb3a567d89b37c06da01459c83d2c54be"} Mar 18 00:16:03 crc kubenswrapper[4983]: I0318 00:16:03.007668 4983 generic.go:334] "Generic (PLEG): container finished" podID="a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf" containerID="1214a03f220209d0bfd44f9dc961ae1622eb933596e325b50d76eb0c01361fb8" exitCode=0 Mar 18 00:16:03 crc kubenswrapper[4983]: I0318 00:16:03.008003 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563216-jj8fm" event={"ID":"a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf","Type":"ContainerDied","Data":"1214a03f220209d0bfd44f9dc961ae1622eb933596e325b50d76eb0c01361fb8"} Mar 18 00:16:04 crc kubenswrapper[4983]: I0318 00:16:04.279692 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563216-jj8fm" Mar 18 00:16:04 crc kubenswrapper[4983]: I0318 00:16:04.340411 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjqwn\" (UniqueName: \"kubernetes.io/projected/a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf-kube-api-access-xjqwn\") pod \"a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf\" (UID: \"a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf\") " Mar 18 00:16:04 crc kubenswrapper[4983]: I0318 00:16:04.345140 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf-kube-api-access-xjqwn" (OuterVolumeSpecName: "kube-api-access-xjqwn") pod "a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf" (UID: "a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf"). InnerVolumeSpecName "kube-api-access-xjqwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:16:04 crc kubenswrapper[4983]: I0318 00:16:04.441143 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjqwn\" (UniqueName: \"kubernetes.io/projected/a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf-kube-api-access-xjqwn\") on node \"crc\" DevicePath \"\"" Mar 18 00:16:05 crc kubenswrapper[4983]: I0318 00:16:05.019442 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563216-jj8fm" event={"ID":"a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf","Type":"ContainerDied","Data":"3a165dbce117b658b4e83612be5864cdb3a567d89b37c06da01459c83d2c54be"} Mar 18 00:16:05 crc kubenswrapper[4983]: I0318 00:16:05.019498 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a165dbce117b658b4e83612be5864cdb3a567d89b37c06da01459c83d2c54be" Mar 18 00:16:05 crc kubenswrapper[4983]: I0318 00:16:05.019523 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563216-jj8fm" Mar 18 00:16:05 crc kubenswrapper[4983]: I0318 00:16:05.354585 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563210-lptd5"] Mar 18 00:16:05 crc kubenswrapper[4983]: I0318 00:16:05.357915 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563210-lptd5"] Mar 18 00:16:06 crc kubenswrapper[4983]: I0318 00:16:06.850608 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efeae5ef-dba6-4031-83be-8f493fe853b9" path="/var/lib/kubelet/pods/efeae5ef-dba6-4031-83be-8f493fe853b9/volumes" Mar 18 00:17:43 crc kubenswrapper[4983]: I0318 00:17:43.259161 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:17:43 crc kubenswrapper[4983]: I0318 00:17:43.260029 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.147617 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563218-f2zls"] Mar 18 00:18:00 crc kubenswrapper[4983]: E0318 00:18:00.148659 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf" containerName="oc" Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.148682 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf" containerName="oc" Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.148880 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf" containerName="oc" Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.149543 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563218-f2zls" Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.153056 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.153632 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.154736 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.170842 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563218-f2zls"] Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.175482 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb9w8\" (UniqueName: \"kubernetes.io/projected/97fb5d36-22f5-493c-a22f-e857b801905c-kube-api-access-qb9w8\") pod \"auto-csr-approver-29563218-f2zls\" (UID: \"97fb5d36-22f5-493c-a22f-e857b801905c\") " pod="openshift-infra/auto-csr-approver-29563218-f2zls" Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.277051 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb9w8\" (UniqueName: \"kubernetes.io/projected/97fb5d36-22f5-493c-a22f-e857b801905c-kube-api-access-qb9w8\") pod \"auto-csr-approver-29563218-f2zls\" (UID: \"97fb5d36-22f5-493c-a22f-e857b801905c\") " pod="openshift-infra/auto-csr-approver-29563218-f2zls" Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.308838 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb9w8\" (UniqueName: \"kubernetes.io/projected/97fb5d36-22f5-493c-a22f-e857b801905c-kube-api-access-qb9w8\") pod \"auto-csr-approver-29563218-f2zls\" (UID: \"97fb5d36-22f5-493c-a22f-e857b801905c\") " pod="openshift-infra/auto-csr-approver-29563218-f2zls" Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.482667 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563218-f2zls" Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.752469 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563218-f2zls"] Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.760443 4983 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 00:18:00 crc kubenswrapper[4983]: I0318 00:18:00.863913 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563218-f2zls" event={"ID":"97fb5d36-22f5-493c-a22f-e857b801905c","Type":"ContainerStarted","Data":"2b24aa505a8b8d454d35e1e751e07726a39d52cc0c824af682b0a0276762254f"} Mar 18 00:18:02 crc kubenswrapper[4983]: I0318 00:18:02.880056 4983 generic.go:334] "Generic (PLEG): container finished" podID="97fb5d36-22f5-493c-a22f-e857b801905c" containerID="89ddb68660edebbbebab0a216bba5801a0590e6e8d8d43fff9058d340c069d97" exitCode=0 Mar 18 00:18:02 crc kubenswrapper[4983]: I0318 00:18:02.880135 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563218-f2zls" event={"ID":"97fb5d36-22f5-493c-a22f-e857b801905c","Type":"ContainerDied","Data":"89ddb68660edebbbebab0a216bba5801a0590e6e8d8d43fff9058d340c069d97"} Mar 18 00:18:04 crc kubenswrapper[4983]: I0318 00:18:04.046446 4983 scope.go:117] "RemoveContainer" containerID="d9eb68e8bcc8979fda5ff8f269ddbb7bc410310bba6eba0fe60bd4b384c28556" Mar 18 00:18:04 crc kubenswrapper[4983]: I0318 00:18:04.081212 4983 scope.go:117] "RemoveContainer" containerID="203afe93d4764619d6478248cfea556a15b41013072fbd9efe114ade133b9a42" Mar 18 00:18:04 crc kubenswrapper[4983]: I0318 00:18:04.109585 4983 scope.go:117] "RemoveContainer" containerID="a7d0b17358ee5ef5868fa4c70dc42c060581f32d2f5170f73dc9e44fb25e5400" Mar 18 00:18:04 crc kubenswrapper[4983]: I0318 00:18:04.162501 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563218-f2zls" Mar 18 00:18:04 crc kubenswrapper[4983]: I0318 00:18:04.238837 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qb9w8\" (UniqueName: \"kubernetes.io/projected/97fb5d36-22f5-493c-a22f-e857b801905c-kube-api-access-qb9w8\") pod \"97fb5d36-22f5-493c-a22f-e857b801905c\" (UID: \"97fb5d36-22f5-493c-a22f-e857b801905c\") " Mar 18 00:18:04 crc kubenswrapper[4983]: I0318 00:18:04.247383 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97fb5d36-22f5-493c-a22f-e857b801905c-kube-api-access-qb9w8" (OuterVolumeSpecName: "kube-api-access-qb9w8") pod "97fb5d36-22f5-493c-a22f-e857b801905c" (UID: "97fb5d36-22f5-493c-a22f-e857b801905c"). InnerVolumeSpecName "kube-api-access-qb9w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:18:04 crc kubenswrapper[4983]: I0318 00:18:04.340777 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qb9w8\" (UniqueName: \"kubernetes.io/projected/97fb5d36-22f5-493c-a22f-e857b801905c-kube-api-access-qb9w8\") on node \"crc\" DevicePath \"\"" Mar 18 00:18:04 crc kubenswrapper[4983]: I0318 00:18:04.898966 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563218-f2zls" event={"ID":"97fb5d36-22f5-493c-a22f-e857b801905c","Type":"ContainerDied","Data":"2b24aa505a8b8d454d35e1e751e07726a39d52cc0c824af682b0a0276762254f"} Mar 18 00:18:04 crc kubenswrapper[4983]: I0318 00:18:04.899016 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563218-f2zls" Mar 18 00:18:04 crc kubenswrapper[4983]: I0318 00:18:04.899036 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b24aa505a8b8d454d35e1e751e07726a39d52cc0c824af682b0a0276762254f" Mar 18 00:18:05 crc kubenswrapper[4983]: I0318 00:18:05.241173 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563212-g7qg2"] Mar 18 00:18:05 crc kubenswrapper[4983]: I0318 00:18:05.248132 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563212-g7qg2"] Mar 18 00:18:06 crc kubenswrapper[4983]: I0318 00:18:06.856850 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55e41d80-a523-4a82-8d42-c01daa50d832" path="/var/lib/kubelet/pods/55e41d80-a523-4a82-8d42-c01daa50d832/volumes" Mar 18 00:18:13 crc kubenswrapper[4983]: I0318 00:18:13.259349 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:18:13 crc kubenswrapper[4983]: I0318 00:18:13.260332 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:18:43 crc kubenswrapper[4983]: I0318 00:18:43.259645 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:18:43 crc kubenswrapper[4983]: I0318 00:18:43.260341 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:18:43 crc kubenswrapper[4983]: I0318 00:18:43.260408 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:18:43 crc kubenswrapper[4983]: I0318 00:18:43.261466 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d469517a2f6224bd41c05c725e6f6ca91ae6c676622fc45a16b3a323fcf16e9"} pod="openshift-machine-config-operator/machine-config-daemon-jczfs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 00:18:43 crc kubenswrapper[4983]: I0318 00:18:43.261558 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" containerID="cri-o://7d469517a2f6224bd41c05c725e6f6ca91ae6c676622fc45a16b3a323fcf16e9" gracePeriod=600 Mar 18 00:18:44 crc kubenswrapper[4983]: I0318 00:18:44.207505 4983 generic.go:334] "Generic (PLEG): container finished" podID="482e2afe-48be-4ea2-8bda-52849c901b52" containerID="7d469517a2f6224bd41c05c725e6f6ca91ae6c676622fc45a16b3a323fcf16e9" exitCode=0 Mar 18 00:18:44 crc kubenswrapper[4983]: I0318 00:18:44.207573 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerDied","Data":"7d469517a2f6224bd41c05c725e6f6ca91ae6c676622fc45a16b3a323fcf16e9"} Mar 18 00:18:44 crc kubenswrapper[4983]: I0318 00:18:44.208050 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerStarted","Data":"9cc35c9c62ec6606ad35039f3a8a4340518e27795998f82a00d733ef5cac33bb"} Mar 18 00:18:44 crc kubenswrapper[4983]: I0318 00:18:44.208086 4983 scope.go:117] "RemoveContainer" containerID="7e5e3b4eea0e0cda0f8f16612d694e8e56064b538ff4f0543482bc9f225a5195" Mar 18 00:19:04 crc kubenswrapper[4983]: I0318 00:19:04.205235 4983 scope.go:117] "RemoveContainer" containerID="2fa0c0dbe2e1b40055a0e3bac3a783bbaace18295265c106a565a4ea2e8a7856" Mar 18 00:19:04 crc kubenswrapper[4983]: I0318 00:19:04.238420 4983 scope.go:117] "RemoveContainer" containerID="c680d35f60a7e80033026aef33a6e7b38bf5ef016a8e826b865081310759914d" Mar 18 00:19:04 crc kubenswrapper[4983]: I0318 00:19:04.296303 4983 scope.go:117] "RemoveContainer" containerID="eda81797bff5f0f42dc4edec65d6455e9985956ab14c064f4e4970b15110ca7f" Mar 18 00:20:00 crc kubenswrapper[4983]: I0318 00:20:00.148698 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563220-zdtdg"] Mar 18 00:20:00 crc kubenswrapper[4983]: E0318 00:20:00.149720 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97fb5d36-22f5-493c-a22f-e857b801905c" containerName="oc" Mar 18 00:20:00 crc kubenswrapper[4983]: I0318 00:20:00.149750 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="97fb5d36-22f5-493c-a22f-e857b801905c" containerName="oc" Mar 18 00:20:00 crc kubenswrapper[4983]: I0318 00:20:00.149968 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="97fb5d36-22f5-493c-a22f-e857b801905c" containerName="oc" Mar 18 00:20:00 crc kubenswrapper[4983]: I0318 00:20:00.150568 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563220-zdtdg" Mar 18 00:20:00 crc kubenswrapper[4983]: I0318 00:20:00.154594 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:20:00 crc kubenswrapper[4983]: I0318 00:20:00.154654 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:20:00 crc kubenswrapper[4983]: I0318 00:20:00.154753 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:20:00 crc kubenswrapper[4983]: I0318 00:20:00.161663 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563220-zdtdg"] Mar 18 00:20:00 crc kubenswrapper[4983]: I0318 00:20:00.282689 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zkpm\" (UniqueName: \"kubernetes.io/projected/d149429e-e5d1-48e5-a052-feecd610f7c5-kube-api-access-7zkpm\") pod \"auto-csr-approver-29563220-zdtdg\" (UID: \"d149429e-e5d1-48e5-a052-feecd610f7c5\") " pod="openshift-infra/auto-csr-approver-29563220-zdtdg" Mar 18 00:20:00 crc kubenswrapper[4983]: I0318 00:20:00.384866 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zkpm\" (UniqueName: \"kubernetes.io/projected/d149429e-e5d1-48e5-a052-feecd610f7c5-kube-api-access-7zkpm\") pod \"auto-csr-approver-29563220-zdtdg\" (UID: \"d149429e-e5d1-48e5-a052-feecd610f7c5\") " pod="openshift-infra/auto-csr-approver-29563220-zdtdg" Mar 18 00:20:00 crc kubenswrapper[4983]: I0318 00:20:00.417998 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zkpm\" (UniqueName: \"kubernetes.io/projected/d149429e-e5d1-48e5-a052-feecd610f7c5-kube-api-access-7zkpm\") pod \"auto-csr-approver-29563220-zdtdg\" (UID: \"d149429e-e5d1-48e5-a052-feecd610f7c5\") " pod="openshift-infra/auto-csr-approver-29563220-zdtdg" Mar 18 00:20:00 crc kubenswrapper[4983]: I0318 00:20:00.515792 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563220-zdtdg" Mar 18 00:20:00 crc kubenswrapper[4983]: I0318 00:20:00.761189 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563220-zdtdg"] Mar 18 00:20:01 crc kubenswrapper[4983]: I0318 00:20:01.755959 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563220-zdtdg" event={"ID":"d149429e-e5d1-48e5-a052-feecd610f7c5","Type":"ContainerStarted","Data":"1604606adbfbb0e09a5c6e94ee6e7797be3cee06401d86c686abf0e98a6c6a60"} Mar 18 00:20:02 crc kubenswrapper[4983]: I0318 00:20:02.764914 4983 generic.go:334] "Generic (PLEG): container finished" podID="d149429e-e5d1-48e5-a052-feecd610f7c5" containerID="41437bb0365646c4a12e203f89ea7a4caa43e57dda1e59a4bb40f5a9e4bae3fe" exitCode=0 Mar 18 00:20:02 crc kubenswrapper[4983]: I0318 00:20:02.765072 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563220-zdtdg" event={"ID":"d149429e-e5d1-48e5-a052-feecd610f7c5","Type":"ContainerDied","Data":"41437bb0365646c4a12e203f89ea7a4caa43e57dda1e59a4bb40f5a9e4bae3fe"} Mar 18 00:20:04 crc kubenswrapper[4983]: I0318 00:20:04.039463 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563220-zdtdg" Mar 18 00:20:04 crc kubenswrapper[4983]: I0318 00:20:04.138176 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zkpm\" (UniqueName: \"kubernetes.io/projected/d149429e-e5d1-48e5-a052-feecd610f7c5-kube-api-access-7zkpm\") pod \"d149429e-e5d1-48e5-a052-feecd610f7c5\" (UID: \"d149429e-e5d1-48e5-a052-feecd610f7c5\") " Mar 18 00:20:04 crc kubenswrapper[4983]: I0318 00:20:04.146523 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d149429e-e5d1-48e5-a052-feecd610f7c5-kube-api-access-7zkpm" (OuterVolumeSpecName: "kube-api-access-7zkpm") pod "d149429e-e5d1-48e5-a052-feecd610f7c5" (UID: "d149429e-e5d1-48e5-a052-feecd610f7c5"). InnerVolumeSpecName "kube-api-access-7zkpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:20:04 crc kubenswrapper[4983]: I0318 00:20:04.240215 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zkpm\" (UniqueName: \"kubernetes.io/projected/d149429e-e5d1-48e5-a052-feecd610f7c5-kube-api-access-7zkpm\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:04 crc kubenswrapper[4983]: I0318 00:20:04.782164 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563220-zdtdg" event={"ID":"d149429e-e5d1-48e5-a052-feecd610f7c5","Type":"ContainerDied","Data":"1604606adbfbb0e09a5c6e94ee6e7797be3cee06401d86c686abf0e98a6c6a60"} Mar 18 00:20:04 crc kubenswrapper[4983]: I0318 00:20:04.782218 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563220-zdtdg" Mar 18 00:20:04 crc kubenswrapper[4983]: I0318 00:20:04.782224 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1604606adbfbb0e09a5c6e94ee6e7797be3cee06401d86c686abf0e98a6c6a60" Mar 18 00:20:05 crc kubenswrapper[4983]: I0318 00:20:05.119869 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563214-v76f9"] Mar 18 00:20:05 crc kubenswrapper[4983]: I0318 00:20:05.126066 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563214-v76f9"] Mar 18 00:20:06 crc kubenswrapper[4983]: I0318 00:20:06.860140 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f97d457f-34d7-4013-9fd5-06cb0cea6720" path="/var/lib/kubelet/pods/f97d457f-34d7-4013-9fd5-06cb0cea6720/volumes" Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.703392 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4r8sk"] Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.704410 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovn-controller" containerID="cri-o://f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39" gracePeriod=30 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.704442 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="sbdb" containerID="cri-o://9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311" gracePeriod=30 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.704547 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18" gracePeriod=30 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.704605 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovn-acl-logging" containerID="cri-o://5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8" gracePeriod=30 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.704605 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="kube-rbac-proxy-node" containerID="cri-o://8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01" gracePeriod=30 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.704753 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="nbdb" containerID="cri-o://1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e" gracePeriod=30 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.704804 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="northd" containerID="cri-o://0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6" gracePeriod=30 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.765646 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" containerID="cri-o://3f11bbbdc7608e518cf73e0bbbff4326a4268dec9707d318c5af979798250551" gracePeriod=30 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.979045 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovnkube-controller/3.log" Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.981594 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovn-acl-logging/0.log" Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982241 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovn-controller/0.log" Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982732 4983 generic.go:334] "Generic (PLEG): container finished" podID="938043fe-450c-4477-8f55-312a843b0706" containerID="3f11bbbdc7608e518cf73e0bbbff4326a4268dec9707d318c5af979798250551" exitCode=0 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982760 4983 generic.go:334] "Generic (PLEG): container finished" podID="938043fe-450c-4477-8f55-312a843b0706" containerID="9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311" exitCode=0 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982769 4983 generic.go:334] "Generic (PLEG): container finished" podID="938043fe-450c-4477-8f55-312a843b0706" containerID="1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e" exitCode=0 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982778 4983 generic.go:334] "Generic (PLEG): container finished" podID="938043fe-450c-4477-8f55-312a843b0706" containerID="0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6" exitCode=0 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982787 4983 generic.go:334] "Generic (PLEG): container finished" podID="938043fe-450c-4477-8f55-312a843b0706" containerID="5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18" exitCode=0 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982795 4983 generic.go:334] "Generic (PLEG): container finished" podID="938043fe-450c-4477-8f55-312a843b0706" containerID="8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01" exitCode=0 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982802 4983 generic.go:334] "Generic (PLEG): container finished" podID="938043fe-450c-4477-8f55-312a843b0706" containerID="5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8" exitCode=143 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982809 4983 generic.go:334] "Generic (PLEG): container finished" podID="938043fe-450c-4477-8f55-312a843b0706" containerID="f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39" exitCode=143 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982818 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"3f11bbbdc7608e518cf73e0bbbff4326a4268dec9707d318c5af979798250551"} Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982878 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311"} Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982891 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e"} Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982902 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6"} Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982913 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18"} Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982938 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01"} Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982950 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8"} Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982956 4983 scope.go:117] "RemoveContainer" containerID="2c9da3016cb68f58bb49fc3fe600ee67021bcf66a99f500126d77a170824d50e" Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.982963 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39"} Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.983067 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" event={"ID":"938043fe-450c-4477-8f55-312a843b0706","Type":"ContainerDied","Data":"b48225f4b6d13dc79a0cb55a03dd7dfcf45305c9e6d68df17c10c35e9b247c5e"} Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.983082 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b48225f4b6d13dc79a0cb55a03dd7dfcf45305c9e6d68df17c10c35e9b247c5e" Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.985071 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6nblt_0853f8de-9bfc-4053-acd0-8893122d2d62/kube-multus/2.log" Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.985526 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6nblt_0853f8de-9bfc-4053-acd0-8893122d2d62/kube-multus/1.log" Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.985563 4983 generic.go:334] "Generic (PLEG): container finished" podID="0853f8de-9bfc-4053-acd0-8893122d2d62" containerID="106e6e771cbb1f79ee805368b6caaefd9fb0ec1fa69685d104bc832621c8f39d" exitCode=2 Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.985613 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6nblt" event={"ID":"0853f8de-9bfc-4053-acd0-8893122d2d62","Type":"ContainerDied","Data":"106e6e771cbb1f79ee805368b6caaefd9fb0ec1fa69685d104bc832621c8f39d"} Mar 18 00:20:31 crc kubenswrapper[4983]: I0318 00:20:31.986238 4983 scope.go:117] "RemoveContainer" containerID="106e6e771cbb1f79ee805368b6caaefd9fb0ec1fa69685d104bc832621c8f39d" Mar 18 00:20:31 crc kubenswrapper[4983]: E0318 00:20:31.986553 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-6nblt_openshift-multus(0853f8de-9bfc-4053-acd0-8893122d2d62)\"" pod="openshift-multus/multus-6nblt" podUID="0853f8de-9bfc-4053-acd0-8893122d2d62" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.020114 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovn-acl-logging/0.log" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.020575 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovn-controller/0.log" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.021042 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.027315 4983 scope.go:117] "RemoveContainer" containerID="96a804156a676d27f5f2f52777505d61cc68bac1341339c979141ceb90b8401b" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077446 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b8g9v"] Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.077643 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovn-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077662 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovn-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.077709 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="nbdb" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077721 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="nbdb" Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.077731 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077739 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.077753 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d149429e-e5d1-48e5-a052-feecd610f7c5" containerName="oc" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077788 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="d149429e-e5d1-48e5-a052-feecd610f7c5" containerName="oc" Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.077799 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="northd" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077807 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="northd" Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.077816 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovn-acl-logging" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077821 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovn-acl-logging" Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.077831 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="sbdb" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077836 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="sbdb" Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.077843 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077868 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.077875 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="kube-rbac-proxy-node" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077881 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="kube-rbac-proxy-node" Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.077887 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077893 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.077900 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077907 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.077914 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="kubecfg-setup" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077920 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="kubecfg-setup" Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.077954 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="kube-rbac-proxy-ovn-metrics" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.077959 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="kube-rbac-proxy-ovn-metrics" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078082 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078110 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="nbdb" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078120 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078129 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="kube-rbac-proxy-ovn-metrics" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078138 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078145 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovn-acl-logging" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078156 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="sbdb" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078163 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="kube-rbac-proxy-node" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078190 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="northd" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078201 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovn-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078209 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="d149429e-e5d1-48e5-a052-feecd610f7c5" containerName="oc" Mar 18 00:20:32 crc kubenswrapper[4983]: E0318 00:20:32.078320 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078346 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078475 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.078661 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="938043fe-450c-4477-8f55-312a843b0706" containerName="ovnkube-controller" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.079999 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.154879 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-kubelet\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155209 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-ovnkube-config\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.154975 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155231 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-openvswitch\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155257 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-log-socket\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155273 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/938043fe-450c-4477-8f55-312a843b0706-ovn-node-metrics-cert\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155290 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-systemd\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155306 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-cni-bin\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155323 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-node-log\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155358 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-slash\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155379 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-var-lib-cni-networks-ovn-kubernetes\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155399 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-env-overrides\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155417 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-var-lib-openvswitch\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155440 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-ovn\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155462 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-run-ovn-kubernetes\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155485 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-cni-netd\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155500 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-etc-openvswitch\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155521 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tj4g\" (UniqueName: \"kubernetes.io/projected/938043fe-450c-4477-8f55-312a843b0706-kube-api-access-6tj4g\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155542 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-ovnkube-script-lib\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155573 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-systemd-units\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155596 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-run-netns\") pod \"938043fe-450c-4477-8f55-312a843b0706\" (UID: \"938043fe-450c-4477-8f55-312a843b0706\") " Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.155810 4983 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156002 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156071 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156117 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156156 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-log-socket" (OuterVolumeSpecName: "log-socket") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156314 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156347 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156366 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156384 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156397 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156490 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-slash" (OuterVolumeSpecName: "host-slash") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156325 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-node-log" (OuterVolumeSpecName: "node-log") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156612 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156819 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156576 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156855 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.156984 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.161736 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/938043fe-450c-4477-8f55-312a843b0706-kube-api-access-6tj4g" (OuterVolumeSpecName: "kube-api-access-6tj4g") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "kube-api-access-6tj4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.162776 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/938043fe-450c-4477-8f55-312a843b0706-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.172838 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "938043fe-450c-4477-8f55-312a843b0706" (UID: "938043fe-450c-4477-8f55-312a843b0706"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257438 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-cni-netd\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257517 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/16982d94-f352-45f1-955c-7841668f7128-ovn-node-metrics-cert\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257586 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/16982d94-f352-45f1-955c-7841668f7128-env-overrides\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257610 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-run-openvswitch\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257673 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-run-ovn\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257692 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-kubelet\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257736 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-node-log\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257764 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-run-ovn-kubernetes\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257787 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-var-lib-openvswitch\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257814 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-systemd-units\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257847 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/16982d94-f352-45f1-955c-7841668f7128-ovnkube-config\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257878 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-etc-openvswitch\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257907 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-run-systemd\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.257963 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-log-socket\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258354 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxfw9\" (UniqueName: \"kubernetes.io/projected/16982d94-f352-45f1-955c-7841668f7128-kube-api-access-rxfw9\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258385 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/16982d94-f352-45f1-955c-7841668f7128-ovnkube-script-lib\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258427 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-slash\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258452 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258479 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-cni-bin\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258502 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-run-netns\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258561 4983 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258577 4983 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258590 4983 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-log-socket\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258604 4983 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/938043fe-450c-4477-8f55-312a843b0706-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258618 4983 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258629 4983 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258640 4983 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-node-log\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258711 4983 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-slash\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258787 4983 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258807 4983 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258824 4983 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258909 4983 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258953 4983 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258965 4983 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258978 4983 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.258991 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tj4g\" (UniqueName: \"kubernetes.io/projected/938043fe-450c-4477-8f55-312a843b0706-kube-api-access-6tj4g\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.259007 4983 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/938043fe-450c-4477-8f55-312a843b0706-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.259021 4983 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.259031 4983 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/938043fe-450c-4477-8f55-312a843b0706-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.359889 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-etc-openvswitch\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.359976 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-run-systemd\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360015 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-log-socket\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360060 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxfw9\" (UniqueName: \"kubernetes.io/projected/16982d94-f352-45f1-955c-7841668f7128-kube-api-access-rxfw9\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360095 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/16982d94-f352-45f1-955c-7841668f7128-ovnkube-script-lib\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360123 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-etc-openvswitch\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360199 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-slash\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360147 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-slash\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360250 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-run-systemd\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360277 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-log-socket\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360367 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360436 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-cni-bin\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360521 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-run-netns\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360587 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360603 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-cni-bin\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360626 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-cni-netd\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360644 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-run-netns\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360721 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/16982d94-f352-45f1-955c-7841668f7128-ovn-node-metrics-cert\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360723 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-cni-netd\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360771 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/16982d94-f352-45f1-955c-7841668f7128-env-overrides\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360804 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-run-openvswitch\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360846 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-run-ovn\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360863 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-run-openvswitch\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360875 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-kubelet\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360910 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-kubelet\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360986 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-run-ovn\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.360996 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-node-log\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.361035 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-node-log\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.361082 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-run-ovn-kubernetes\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.361140 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-var-lib-openvswitch\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.361194 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-systemd-units\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.361268 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/16982d94-f352-45f1-955c-7841668f7128-ovnkube-config\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.361552 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/16982d94-f352-45f1-955c-7841668f7128-env-overrides\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.361638 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-var-lib-openvswitch\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.361705 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-host-run-ovn-kubernetes\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.361768 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/16982d94-f352-45f1-955c-7841668f7128-systemd-units\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.362631 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/16982d94-f352-45f1-955c-7841668f7128-ovnkube-script-lib\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.362924 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/16982d94-f352-45f1-955c-7841668f7128-ovnkube-config\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.366661 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/16982d94-f352-45f1-955c-7841668f7128-ovn-node-metrics-cert\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.389333 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxfw9\" (UniqueName: \"kubernetes.io/projected/16982d94-f352-45f1-955c-7841668f7128-kube-api-access-rxfw9\") pod \"ovnkube-node-b8g9v\" (UID: \"16982d94-f352-45f1-955c-7841668f7128\") " pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.396649 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.993706 4983 generic.go:334] "Generic (PLEG): container finished" podID="16982d94-f352-45f1-955c-7841668f7128" containerID="71a028a94552bdb20ecc940a8b5fafffe49be42525b090b68bbd51ed0e2db87d" exitCode=0 Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.993800 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" event={"ID":"16982d94-f352-45f1-955c-7841668f7128","Type":"ContainerDied","Data":"71a028a94552bdb20ecc940a8b5fafffe49be42525b090b68bbd51ed0e2db87d"} Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.994280 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" event={"ID":"16982d94-f352-45f1-955c-7841668f7128","Type":"ContainerStarted","Data":"d9eff39134d2d88f9d6941f29ee1f953334046db40e70136ac379b408340b9ad"} Mar 18 00:20:32 crc kubenswrapper[4983]: I0318 00:20:32.997964 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6nblt_0853f8de-9bfc-4053-acd0-8893122d2d62/kube-multus/2.log" Mar 18 00:20:33 crc kubenswrapper[4983]: I0318 00:20:33.003240 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovn-acl-logging/0.log" Mar 18 00:20:33 crc kubenswrapper[4983]: I0318 00:20:33.003844 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4r8sk_938043fe-450c-4477-8f55-312a843b0706/ovn-controller/0.log" Mar 18 00:20:33 crc kubenswrapper[4983]: I0318 00:20:33.004548 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4r8sk" Mar 18 00:20:33 crc kubenswrapper[4983]: I0318 00:20:33.091095 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4r8sk"] Mar 18 00:20:33 crc kubenswrapper[4983]: I0318 00:20:33.103358 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4r8sk"] Mar 18 00:20:34 crc kubenswrapper[4983]: I0318 00:20:34.018653 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" event={"ID":"16982d94-f352-45f1-955c-7841668f7128","Type":"ContainerStarted","Data":"b9c8fe32469c71d304040090cac61d0a07e246e563ec9bc3f50d55ef70858763"} Mar 18 00:20:34 crc kubenswrapper[4983]: I0318 00:20:34.019103 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" event={"ID":"16982d94-f352-45f1-955c-7841668f7128","Type":"ContainerStarted","Data":"f07cdd3e2c88bca24341b49698ceaa162f808965d77c3f4f45e477459e4c85cd"} Mar 18 00:20:34 crc kubenswrapper[4983]: I0318 00:20:34.019130 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" event={"ID":"16982d94-f352-45f1-955c-7841668f7128","Type":"ContainerStarted","Data":"f8649c070c88c5a46472d4be9fcae9f6d43c6746edaed7a12078963e4d0dfbaa"} Mar 18 00:20:34 crc kubenswrapper[4983]: I0318 00:20:34.019150 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" event={"ID":"16982d94-f352-45f1-955c-7841668f7128","Type":"ContainerStarted","Data":"3f25da1f53dd50d45205f3f64a66acf0114fe82ace5f21056fa6e24b8afb1022"} Mar 18 00:20:34 crc kubenswrapper[4983]: I0318 00:20:34.019167 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" event={"ID":"16982d94-f352-45f1-955c-7841668f7128","Type":"ContainerStarted","Data":"bfd10d59aac1e407d714b33f8b3e28c09ed12aff6ad419f2b4a8a4a5a1076f4e"} Mar 18 00:20:34 crc kubenswrapper[4983]: I0318 00:20:34.019184 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" event={"ID":"16982d94-f352-45f1-955c-7841668f7128","Type":"ContainerStarted","Data":"885828b378f5f27a355316e2ff0230069df4f27375b8b7f2ac006c44cc46a988"} Mar 18 00:20:34 crc kubenswrapper[4983]: I0318 00:20:34.858064 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="938043fe-450c-4477-8f55-312a843b0706" path="/var/lib/kubelet/pods/938043fe-450c-4477-8f55-312a843b0706/volumes" Mar 18 00:20:36 crc kubenswrapper[4983]: I0318 00:20:36.042081 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" event={"ID":"16982d94-f352-45f1-955c-7841668f7128","Type":"ContainerStarted","Data":"8771a4ad2c36753e97c26b4f29800e5c8c9b48f94ba746d22515b7b24abb382e"} Mar 18 00:20:39 crc kubenswrapper[4983]: I0318 00:20:39.089092 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" event={"ID":"16982d94-f352-45f1-955c-7841668f7128","Type":"ContainerStarted","Data":"b2bce542c294d649719b6a2d4c4343eb85abcf6dd3fee6b80e3dd504326768b6"} Mar 18 00:20:39 crc kubenswrapper[4983]: I0318 00:20:39.089706 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:39 crc kubenswrapper[4983]: I0318 00:20:39.090293 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:39 crc kubenswrapper[4983]: I0318 00:20:39.090336 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:39 crc kubenswrapper[4983]: I0318 00:20:39.139113 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" podStartSLOduration=7.139094082 podStartE2EDuration="7.139094082s" podCreationTimestamp="2026-03-18 00:20:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:20:39.135884911 +0000 UTC m=+757.033259581" watchObservedRunningTime="2026-03-18 00:20:39.139094082 +0000 UTC m=+757.036468702" Mar 18 00:20:39 crc kubenswrapper[4983]: I0318 00:20:39.146489 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:39 crc kubenswrapper[4983]: I0318 00:20:39.148796 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:20:43 crc kubenswrapper[4983]: I0318 00:20:43.259204 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:20:43 crc kubenswrapper[4983]: I0318 00:20:43.259944 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:20:46 crc kubenswrapper[4983]: I0318 00:20:46.844876 4983 scope.go:117] "RemoveContainer" containerID="106e6e771cbb1f79ee805368b6caaefd9fb0ec1fa69685d104bc832621c8f39d" Mar 18 00:20:46 crc kubenswrapper[4983]: E0318 00:20:46.845664 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-6nblt_openshift-multus(0853f8de-9bfc-4053-acd0-8893122d2d62)\"" pod="openshift-multus/multus-6nblt" podUID="0853f8de-9bfc-4053-acd0-8893122d2d62" Mar 18 00:20:57 crc kubenswrapper[4983]: I0318 00:20:57.844336 4983 scope.go:117] "RemoveContainer" containerID="106e6e771cbb1f79ee805368b6caaefd9fb0ec1fa69685d104bc832621c8f39d" Mar 18 00:20:58 crc kubenswrapper[4983]: I0318 00:20:58.229048 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6nblt_0853f8de-9bfc-4053-acd0-8893122d2d62/kube-multus/2.log" Mar 18 00:20:58 crc kubenswrapper[4983]: I0318 00:20:58.229522 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6nblt" event={"ID":"0853f8de-9bfc-4053-acd0-8893122d2d62","Type":"ContainerStarted","Data":"1d5c562b30a06aaf07c9db0977c5b7461f0652ea1054c807612665993544ed79"} Mar 18 00:21:02 crc kubenswrapper[4983]: I0318 00:21:02.446026 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b8g9v" Mar 18 00:21:04 crc kubenswrapper[4983]: I0318 00:21:04.376679 4983 scope.go:117] "RemoveContainer" containerID="5e0d4c7ff727e24c02a4fbb0236f3ca1ddac33d347c200d81a703799ef056bd9" Mar 18 00:21:04 crc kubenswrapper[4983]: I0318 00:21:04.406580 4983 scope.go:117] "RemoveContainer" containerID="1f47545948b7f643fb1b169580594693231ebc2ccfd8eee15fd42db9d5867c5e" Mar 18 00:21:04 crc kubenswrapper[4983]: I0318 00:21:04.439836 4983 scope.go:117] "RemoveContainer" containerID="5696d6567ccacd4be1c6ce3e8e7db1580c262a056c116d4fd3181d0b827412b8" Mar 18 00:21:04 crc kubenswrapper[4983]: I0318 00:21:04.464067 4983 scope.go:117] "RemoveContainer" containerID="f600cfda114e5a6574eff91b339af10fdbaf47d0b2edb09e0d6f962bd08dbc36" Mar 18 00:21:04 crc kubenswrapper[4983]: I0318 00:21:04.513083 4983 scope.go:117] "RemoveContainer" containerID="9cf631b92bfa0513462c66acd2e82a27eb828c74a3753662aa63281f43023311" Mar 18 00:21:04 crc kubenswrapper[4983]: I0318 00:21:04.546824 4983 scope.go:117] "RemoveContainer" containerID="0b14f4b997c7a96f73a87ddbf300927104fa837827515c4bc48340d05f454be6" Mar 18 00:21:04 crc kubenswrapper[4983]: I0318 00:21:04.571082 4983 scope.go:117] "RemoveContainer" containerID="5d5a3bef179a09dbb5711c93c5105f87c829129bc156c27d64b4fe433203fd18" Mar 18 00:21:04 crc kubenswrapper[4983]: I0318 00:21:04.593096 4983 scope.go:117] "RemoveContainer" containerID="f26ab9e9049069732d9c0a6e6e8655f5cadfc31c56cf11640275ad297c45ab39" Mar 18 00:21:04 crc kubenswrapper[4983]: I0318 00:21:04.616335 4983 scope.go:117] "RemoveContainer" containerID="8d5cfaeb553676c285e8c3d504d32682f95bd349f007ca7555fd75602a167c01" Mar 18 00:21:04 crc kubenswrapper[4983]: I0318 00:21:04.642060 4983 scope.go:117] "RemoveContainer" containerID="3f11bbbdc7608e518cf73e0bbbff4326a4268dec9707d318c5af979798250551" Mar 18 00:21:13 crc kubenswrapper[4983]: I0318 00:21:13.259546 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:21:13 crc kubenswrapper[4983]: I0318 00:21:13.260144 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:21:29 crc kubenswrapper[4983]: I0318 00:21:29.849041 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6q45q"] Mar 18 00:21:29 crc kubenswrapper[4983]: I0318 00:21:29.850198 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6q45q" podUID="55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" containerName="registry-server" containerID="cri-o://d0f900680a081dabf19a562ca3dfce551f367d448b3d71252a6134f65cedb225" gracePeriod=30 Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.311188 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.465980 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-catalog-content\") pod \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\" (UID: \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\") " Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.466046 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rpjb\" (UniqueName: \"kubernetes.io/projected/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-kube-api-access-6rpjb\") pod \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\" (UID: \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\") " Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.466125 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-utilities\") pod \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\" (UID: \"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb\") " Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.467073 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-utilities" (OuterVolumeSpecName: "utilities") pod "55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" (UID: "55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.473576 4983 generic.go:334] "Generic (PLEG): container finished" podID="55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" containerID="d0f900680a081dabf19a562ca3dfce551f367d448b3d71252a6134f65cedb225" exitCode=0 Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.473627 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6q45q" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.473626 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6q45q" event={"ID":"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb","Type":"ContainerDied","Data":"d0f900680a081dabf19a562ca3dfce551f367d448b3d71252a6134f65cedb225"} Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.473707 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6q45q" event={"ID":"55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb","Type":"ContainerDied","Data":"e6dfd2ffdbb2e56d91a1683388fd4f452da2fa0c1a7d969762a1028d64a13eae"} Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.473743 4983 scope.go:117] "RemoveContainer" containerID="d0f900680a081dabf19a562ca3dfce551f367d448b3d71252a6134f65cedb225" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.475950 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-kube-api-access-6rpjb" (OuterVolumeSpecName: "kube-api-access-6rpjb") pod "55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" (UID: "55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb"). InnerVolumeSpecName "kube-api-access-6rpjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.519291 4983 scope.go:117] "RemoveContainer" containerID="d9643f6c4875b52c89c8525ecb86ec72bc696d7ea7167509255edcf1120effc2" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.520223 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" (UID: "55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.538500 4983 scope.go:117] "RemoveContainer" containerID="7b474c7667f4e39bceb7398e335a681e7f66a995c0364d8043ae0d34d174c346" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.569357 4983 scope.go:117] "RemoveContainer" containerID="d0f900680a081dabf19a562ca3dfce551f367d448b3d71252a6134f65cedb225" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.569961 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.570028 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.570055 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rpjb\" (UniqueName: \"kubernetes.io/projected/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb-kube-api-access-6rpjb\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:30 crc kubenswrapper[4983]: E0318 00:21:30.570149 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0f900680a081dabf19a562ca3dfce551f367d448b3d71252a6134f65cedb225\": container with ID starting with d0f900680a081dabf19a562ca3dfce551f367d448b3d71252a6134f65cedb225 not found: ID does not exist" containerID="d0f900680a081dabf19a562ca3dfce551f367d448b3d71252a6134f65cedb225" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.570265 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0f900680a081dabf19a562ca3dfce551f367d448b3d71252a6134f65cedb225"} err="failed to get container status \"d0f900680a081dabf19a562ca3dfce551f367d448b3d71252a6134f65cedb225\": rpc error: code = NotFound desc = could not find container \"d0f900680a081dabf19a562ca3dfce551f367d448b3d71252a6134f65cedb225\": container with ID starting with d0f900680a081dabf19a562ca3dfce551f367d448b3d71252a6134f65cedb225 not found: ID does not exist" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.570348 4983 scope.go:117] "RemoveContainer" containerID="d9643f6c4875b52c89c8525ecb86ec72bc696d7ea7167509255edcf1120effc2" Mar 18 00:21:30 crc kubenswrapper[4983]: E0318 00:21:30.570850 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9643f6c4875b52c89c8525ecb86ec72bc696d7ea7167509255edcf1120effc2\": container with ID starting with d9643f6c4875b52c89c8525ecb86ec72bc696d7ea7167509255edcf1120effc2 not found: ID does not exist" containerID="d9643f6c4875b52c89c8525ecb86ec72bc696d7ea7167509255edcf1120effc2" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.570896 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9643f6c4875b52c89c8525ecb86ec72bc696d7ea7167509255edcf1120effc2"} err="failed to get container status \"d9643f6c4875b52c89c8525ecb86ec72bc696d7ea7167509255edcf1120effc2\": rpc error: code = NotFound desc = could not find container \"d9643f6c4875b52c89c8525ecb86ec72bc696d7ea7167509255edcf1120effc2\": container with ID starting with d9643f6c4875b52c89c8525ecb86ec72bc696d7ea7167509255edcf1120effc2 not found: ID does not exist" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.570953 4983 scope.go:117] "RemoveContainer" containerID="7b474c7667f4e39bceb7398e335a681e7f66a995c0364d8043ae0d34d174c346" Mar 18 00:21:30 crc kubenswrapper[4983]: E0318 00:21:30.571393 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b474c7667f4e39bceb7398e335a681e7f66a995c0364d8043ae0d34d174c346\": container with ID starting with 7b474c7667f4e39bceb7398e335a681e7f66a995c0364d8043ae0d34d174c346 not found: ID does not exist" containerID="7b474c7667f4e39bceb7398e335a681e7f66a995c0364d8043ae0d34d174c346" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.571421 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b474c7667f4e39bceb7398e335a681e7f66a995c0364d8043ae0d34d174c346"} err="failed to get container status \"7b474c7667f4e39bceb7398e335a681e7f66a995c0364d8043ae0d34d174c346\": rpc error: code = NotFound desc = could not find container \"7b474c7667f4e39bceb7398e335a681e7f66a995c0364d8043ae0d34d174c346\": container with ID starting with 7b474c7667f4e39bceb7398e335a681e7f66a995c0364d8043ae0d34d174c346 not found: ID does not exist" Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.801344 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6q45q"] Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.805612 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6q45q"] Mar 18 00:21:30 crc kubenswrapper[4983]: I0318 00:21:30.851102 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" path="/var/lib/kubelet/pods/55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb/volumes" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.024571 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm"] Mar 18 00:21:34 crc kubenswrapper[4983]: E0318 00:21:34.025172 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" containerName="extract-content" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.025193 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" containerName="extract-content" Mar 18 00:21:34 crc kubenswrapper[4983]: E0318 00:21:34.025224 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" containerName="extract-utilities" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.025237 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" containerName="extract-utilities" Mar 18 00:21:34 crc kubenswrapper[4983]: E0318 00:21:34.025263 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" containerName="registry-server" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.025275 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" containerName="registry-server" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.025422 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="55b8f13f-d4a3-4efb-a1b1-2e8d44aaaafb" containerName="registry-server" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.026588 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.030039 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.046636 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm"] Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.109980 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12a26068-f547-44fb-b891-b43a072f3667-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm\" (UID: \"12a26068-f547-44fb-b891-b43a072f3667\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.110070 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12a26068-f547-44fb-b891-b43a072f3667-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm\" (UID: \"12a26068-f547-44fb-b891-b43a072f3667\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.110158 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7zdm\" (UniqueName: \"kubernetes.io/projected/12a26068-f547-44fb-b891-b43a072f3667-kube-api-access-z7zdm\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm\" (UID: \"12a26068-f547-44fb-b891-b43a072f3667\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.211037 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12a26068-f547-44fb-b891-b43a072f3667-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm\" (UID: \"12a26068-f547-44fb-b891-b43a072f3667\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.211116 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12a26068-f547-44fb-b891-b43a072f3667-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm\" (UID: \"12a26068-f547-44fb-b891-b43a072f3667\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.211175 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7zdm\" (UniqueName: \"kubernetes.io/projected/12a26068-f547-44fb-b891-b43a072f3667-kube-api-access-z7zdm\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm\" (UID: \"12a26068-f547-44fb-b891-b43a072f3667\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.211849 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12a26068-f547-44fb-b891-b43a072f3667-bundle\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm\" (UID: \"12a26068-f547-44fb-b891-b43a072f3667\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.212111 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12a26068-f547-44fb-b891-b43a072f3667-util\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm\" (UID: \"12a26068-f547-44fb-b891-b43a072f3667\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.233687 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7zdm\" (UniqueName: \"kubernetes.io/projected/12a26068-f547-44fb-b891-b43a072f3667-kube-api-access-z7zdm\") pod \"93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm\" (UID: \"12a26068-f547-44fb-b891-b43a072f3667\") " pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.358874 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" Mar 18 00:21:34 crc kubenswrapper[4983]: I0318 00:21:34.589639 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm"] Mar 18 00:21:35 crc kubenswrapper[4983]: I0318 00:21:35.503959 4983 generic.go:334] "Generic (PLEG): container finished" podID="12a26068-f547-44fb-b891-b43a072f3667" containerID="5afd77e27c50e5b8598a5793a4bc8e98bd79be7c57b0394b0a805473bc6aa606" exitCode=0 Mar 18 00:21:35 crc kubenswrapper[4983]: I0318 00:21:35.504010 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" event={"ID":"12a26068-f547-44fb-b891-b43a072f3667","Type":"ContainerDied","Data":"5afd77e27c50e5b8598a5793a4bc8e98bd79be7c57b0394b0a805473bc6aa606"} Mar 18 00:21:35 crc kubenswrapper[4983]: I0318 00:21:35.504043 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" event={"ID":"12a26068-f547-44fb-b891-b43a072f3667","Type":"ContainerStarted","Data":"f03237e7c31b46e16caa778c1130acfd9c36fc697146fba4c29dccd7ee81dc2f"} Mar 18 00:21:37 crc kubenswrapper[4983]: I0318 00:21:37.523564 4983 generic.go:334] "Generic (PLEG): container finished" podID="12a26068-f547-44fb-b891-b43a072f3667" containerID="bd56d590b8b571719f3665c801845885fbc52f5ad203cf6b7a5678fa98b3b0f3" exitCode=0 Mar 18 00:21:37 crc kubenswrapper[4983]: I0318 00:21:37.523615 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" event={"ID":"12a26068-f547-44fb-b891-b43a072f3667","Type":"ContainerDied","Data":"bd56d590b8b571719f3665c801845885fbc52f5ad203cf6b7a5678fa98b3b0f3"} Mar 18 00:21:38 crc kubenswrapper[4983]: I0318 00:21:38.533408 4983 generic.go:334] "Generic (PLEG): container finished" podID="12a26068-f547-44fb-b891-b43a072f3667" containerID="1d927dfefc02a9a30f454c2b45be290710a8102c5a906b0b65248afbb4181cf3" exitCode=0 Mar 18 00:21:38 crc kubenswrapper[4983]: I0318 00:21:38.533474 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" event={"ID":"12a26068-f547-44fb-b891-b43a072f3667","Type":"ContainerDied","Data":"1d927dfefc02a9a30f454c2b45be290710a8102c5a906b0b65248afbb4181cf3"} Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.199525 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4"] Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.201745 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.210552 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4"] Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.297924 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g28lg\" (UniqueName: \"kubernetes.io/projected/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-kube-api-access-g28lg\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4\" (UID: \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.298017 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4\" (UID: \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.298240 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4\" (UID: \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.399788 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g28lg\" (UniqueName: \"kubernetes.io/projected/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-kube-api-access-g28lg\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4\" (UID: \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.399862 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4\" (UID: \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.400041 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4\" (UID: \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.400749 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4\" (UID: \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.400891 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4\" (UID: \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.438042 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g28lg\" (UniqueName: \"kubernetes.io/projected/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-kube-api-access-g28lg\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4\" (UID: \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.529753 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" Mar 18 00:21:39 crc kubenswrapper[4983]: I0318 00:21:39.831269 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.002295 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4"] Mar 18 00:21:40 crc kubenswrapper[4983]: W0318 00:21:40.004890 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d28b0cd_21c7_4ed7_ab99_5f8e7a7113a2.slice/crio-9482255574ed9e1b6ab76d94774513b5053c6f636408bbead574f83c68ca82a0 WatchSource:0}: Error finding container 9482255574ed9e1b6ab76d94774513b5053c6f636408bbead574f83c68ca82a0: Status 404 returned error can't find the container with id 9482255574ed9e1b6ab76d94774513b5053c6f636408bbead574f83c68ca82a0 Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.005770 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7zdm\" (UniqueName: \"kubernetes.io/projected/12a26068-f547-44fb-b891-b43a072f3667-kube-api-access-z7zdm\") pod \"12a26068-f547-44fb-b891-b43a072f3667\" (UID: \"12a26068-f547-44fb-b891-b43a072f3667\") " Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.005845 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12a26068-f547-44fb-b891-b43a072f3667-bundle\") pod \"12a26068-f547-44fb-b891-b43a072f3667\" (UID: \"12a26068-f547-44fb-b891-b43a072f3667\") " Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.009646 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12a26068-f547-44fb-b891-b43a072f3667-util\") pod \"12a26068-f547-44fb-b891-b43a072f3667\" (UID: \"12a26068-f547-44fb-b891-b43a072f3667\") " Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.012109 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12a26068-f547-44fb-b891-b43a072f3667-bundle" (OuterVolumeSpecName: "bundle") pod "12a26068-f547-44fb-b891-b43a072f3667" (UID: "12a26068-f547-44fb-b891-b43a072f3667"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.012916 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl"] Mar 18 00:21:40 crc kubenswrapper[4983]: E0318 00:21:40.013812 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a26068-f547-44fb-b891-b43a072f3667" containerName="pull" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.013913 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a26068-f547-44fb-b891-b43a072f3667" containerName="pull" Mar 18 00:21:40 crc kubenswrapper[4983]: E0318 00:21:40.013949 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a26068-f547-44fb-b891-b43a072f3667" containerName="util" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.013974 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a26068-f547-44fb-b891-b43a072f3667" containerName="util" Mar 18 00:21:40 crc kubenswrapper[4983]: E0318 00:21:40.013988 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a26068-f547-44fb-b891-b43a072f3667" containerName="extract" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.013996 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a26068-f547-44fb-b891-b43a072f3667" containerName="extract" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.014223 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12a26068-f547-44fb-b891-b43a072f3667-kube-api-access-z7zdm" (OuterVolumeSpecName: "kube-api-access-z7zdm") pod "12a26068-f547-44fb-b891-b43a072f3667" (UID: "12a26068-f547-44fb-b891-b43a072f3667"). InnerVolumeSpecName "kube-api-access-z7zdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.014546 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="12a26068-f547-44fb-b891-b43a072f3667" containerName="extract" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.016641 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.020511 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl"] Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.039422 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12a26068-f547-44fb-b891-b43a072f3667-util" (OuterVolumeSpecName: "util") pod "12a26068-f547-44fb-b891-b43a072f3667" (UID: "12a26068-f547-44fb-b891-b43a072f3667"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.112238 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl\" (UID: \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.112277 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl\" (UID: \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.112306 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2jlb\" (UniqueName: \"kubernetes.io/projected/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-kube-api-access-j2jlb\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl\" (UID: \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.112370 4983 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12a26068-f547-44fb-b891-b43a072f3667-util\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.112383 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7zdm\" (UniqueName: \"kubernetes.io/projected/12a26068-f547-44fb-b891-b43a072f3667-kube-api-access-z7zdm\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.112393 4983 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12a26068-f547-44fb-b891-b43a072f3667-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.213533 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl\" (UID: \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.213986 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl\" (UID: \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.214043 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl\" (UID: \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.214046 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2jlb\" (UniqueName: \"kubernetes.io/projected/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-kube-api-access-j2jlb\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl\" (UID: \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.214567 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl\" (UID: \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.245654 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2jlb\" (UniqueName: \"kubernetes.io/projected/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-kube-api-access-j2jlb\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl\" (UID: \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.346547 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.550558 4983 generic.go:334] "Generic (PLEG): container finished" podID="5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2" containerID="fbc2d2c859248c7be65cfe20f6b12829f1a5568c22732f77d227620d96698f3a" exitCode=0 Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.550639 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" event={"ID":"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2","Type":"ContainerDied","Data":"fbc2d2c859248c7be65cfe20f6b12829f1a5568c22732f77d227620d96698f3a"} Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.550670 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" event={"ID":"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2","Type":"ContainerStarted","Data":"9482255574ed9e1b6ab76d94774513b5053c6f636408bbead574f83c68ca82a0"} Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.554517 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" event={"ID":"12a26068-f547-44fb-b891-b43a072f3667","Type":"ContainerDied","Data":"f03237e7c31b46e16caa778c1130acfd9c36fc697146fba4c29dccd7ee81dc2f"} Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.554572 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f03237e7c31b46e16caa778c1130acfd9c36fc697146fba4c29dccd7ee81dc2f" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.554811 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm" Mar 18 00:21:40 crc kubenswrapper[4983]: I0318 00:21:40.873133 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl"] Mar 18 00:21:40 crc kubenswrapper[4983]: W0318 00:21:40.890694 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb75a947d_18e9_412c_b3d0_2b5e6dcdfed4.slice/crio-9017d2caf50666bef96d10c20955944a3b903639685ec60563e8d56d8f6c2c73 WatchSource:0}: Error finding container 9017d2caf50666bef96d10c20955944a3b903639685ec60563e8d56d8f6c2c73: Status 404 returned error can't find the container with id 9017d2caf50666bef96d10c20955944a3b903639685ec60563e8d56d8f6c2c73 Mar 18 00:21:41 crc kubenswrapper[4983]: I0318 00:21:41.563983 4983 generic.go:334] "Generic (PLEG): container finished" podID="b75a947d-18e9-412c-b3d0-2b5e6dcdfed4" containerID="9c80133eb638201c28f75bd7058adcdd2be0e2ebee4258ddf257c3921446381e" exitCode=0 Mar 18 00:21:41 crc kubenswrapper[4983]: I0318 00:21:41.564168 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" event={"ID":"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4","Type":"ContainerDied","Data":"9c80133eb638201c28f75bd7058adcdd2be0e2ebee4258ddf257c3921446381e"} Mar 18 00:21:41 crc kubenswrapper[4983]: I0318 00:21:41.564481 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" event={"ID":"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4","Type":"ContainerStarted","Data":"9017d2caf50666bef96d10c20955944a3b903639685ec60563e8d56d8f6c2c73"} Mar 18 00:21:42 crc kubenswrapper[4983]: I0318 00:21:42.572815 4983 generic.go:334] "Generic (PLEG): container finished" podID="5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2" containerID="7d6d758516d0719bcf9d9e3e727b1add0bc04ab94211b70b56b5be12817f89ec" exitCode=0 Mar 18 00:21:42 crc kubenswrapper[4983]: I0318 00:21:42.573174 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" event={"ID":"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2","Type":"ContainerDied","Data":"7d6d758516d0719bcf9d9e3e727b1add0bc04ab94211b70b56b5be12817f89ec"} Mar 18 00:21:42 crc kubenswrapper[4983]: I0318 00:21:42.575167 4983 generic.go:334] "Generic (PLEG): container finished" podID="b75a947d-18e9-412c-b3d0-2b5e6dcdfed4" containerID="ebfca746745d7179c86ac32d8d92553d178a251a234d2740031fd9693bef87da" exitCode=0 Mar 18 00:21:42 crc kubenswrapper[4983]: I0318 00:21:42.575233 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" event={"ID":"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4","Type":"ContainerDied","Data":"ebfca746745d7179c86ac32d8d92553d178a251a234d2740031fd9693bef87da"} Mar 18 00:21:43 crc kubenswrapper[4983]: I0318 00:21:43.259755 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:21:43 crc kubenswrapper[4983]: I0318 00:21:43.260058 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:21:43 crc kubenswrapper[4983]: I0318 00:21:43.260112 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:21:43 crc kubenswrapper[4983]: I0318 00:21:43.260799 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9cc35c9c62ec6606ad35039f3a8a4340518e27795998f82a00d733ef5cac33bb"} pod="openshift-machine-config-operator/machine-config-daemon-jczfs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 00:21:43 crc kubenswrapper[4983]: I0318 00:21:43.260866 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" containerID="cri-o://9cc35c9c62ec6606ad35039f3a8a4340518e27795998f82a00d733ef5cac33bb" gracePeriod=600 Mar 18 00:21:43 crc kubenswrapper[4983]: I0318 00:21:43.582577 4983 generic.go:334] "Generic (PLEG): container finished" podID="482e2afe-48be-4ea2-8bda-52849c901b52" containerID="9cc35c9c62ec6606ad35039f3a8a4340518e27795998f82a00d733ef5cac33bb" exitCode=0 Mar 18 00:21:43 crc kubenswrapper[4983]: I0318 00:21:43.582655 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerDied","Data":"9cc35c9c62ec6606ad35039f3a8a4340518e27795998f82a00d733ef5cac33bb"} Mar 18 00:21:43 crc kubenswrapper[4983]: I0318 00:21:43.582708 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerStarted","Data":"81d24af741ccda81f14aaade4f3fc2f7c8569f253f97a8c8f9b6fa245fbd1638"} Mar 18 00:21:43 crc kubenswrapper[4983]: I0318 00:21:43.582731 4983 scope.go:117] "RemoveContainer" containerID="7d469517a2f6224bd41c05c725e6f6ca91ae6c676622fc45a16b3a323fcf16e9" Mar 18 00:21:43 crc kubenswrapper[4983]: I0318 00:21:43.584877 4983 generic.go:334] "Generic (PLEG): container finished" podID="5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2" containerID="49de415559802bd02fff47130dde90acb76fff28e718a849daf145c03b57366b" exitCode=0 Mar 18 00:21:43 crc kubenswrapper[4983]: I0318 00:21:43.584939 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" event={"ID":"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2","Type":"ContainerDied","Data":"49de415559802bd02fff47130dde90acb76fff28e718a849daf145c03b57366b"} Mar 18 00:21:43 crc kubenswrapper[4983]: I0318 00:21:43.586767 4983 generic.go:334] "Generic (PLEG): container finished" podID="b75a947d-18e9-412c-b3d0-2b5e6dcdfed4" containerID="1225f2e5ffdcf53ee9a8020f97c6ccc5e74727c91c6fa75cbd6f50429fb689d9" exitCode=0 Mar 18 00:21:43 crc kubenswrapper[4983]: I0318 00:21:43.586828 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" event={"ID":"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4","Type":"ContainerDied","Data":"1225f2e5ffdcf53ee9a8020f97c6ccc5e74727c91c6fa75cbd6f50429fb689d9"} Mar 18 00:21:44 crc kubenswrapper[4983]: I0318 00:21:44.956703 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" Mar 18 00:21:44 crc kubenswrapper[4983]: I0318 00:21:44.980726 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-util\") pod \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\" (UID: \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\") " Mar 18 00:21:44 crc kubenswrapper[4983]: I0318 00:21:44.980772 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-bundle\") pod \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\" (UID: \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\") " Mar 18 00:21:44 crc kubenswrapper[4983]: I0318 00:21:44.980828 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g28lg\" (UniqueName: \"kubernetes.io/projected/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-kube-api-access-g28lg\") pod \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\" (UID: \"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2\") " Mar 18 00:21:44 crc kubenswrapper[4983]: I0318 00:21:44.982287 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-bundle" (OuterVolumeSpecName: "bundle") pod "5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2" (UID: "5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:21:44 crc kubenswrapper[4983]: I0318 00:21:44.999313 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-util" (OuterVolumeSpecName: "util") pod "5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2" (UID: "5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.001102 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-kube-api-access-g28lg" (OuterVolumeSpecName: "kube-api-access-g28lg") pod "5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2" (UID: "5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2"). InnerVolumeSpecName "kube-api-access-g28lg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.067660 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.083520 4983 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-util\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.083549 4983 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.083562 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g28lg\" (UniqueName: \"kubernetes.io/projected/5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2-kube-api-access-g28lg\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.184769 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-bundle\") pod \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\" (UID: \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\") " Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.184839 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-util\") pod \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\" (UID: \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\") " Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.184870 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2jlb\" (UniqueName: \"kubernetes.io/projected/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-kube-api-access-j2jlb\") pod \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\" (UID: \"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4\") " Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.186230 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-bundle" (OuterVolumeSpecName: "bundle") pod "b75a947d-18e9-412c-b3d0-2b5e6dcdfed4" (UID: "b75a947d-18e9-412c-b3d0-2b5e6dcdfed4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.188793 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-kube-api-access-j2jlb" (OuterVolumeSpecName: "kube-api-access-j2jlb") pod "b75a947d-18e9-412c-b3d0-2b5e6dcdfed4" (UID: "b75a947d-18e9-412c-b3d0-2b5e6dcdfed4"). InnerVolumeSpecName "kube-api-access-j2jlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.202478 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-util" (OuterVolumeSpecName: "util") pod "b75a947d-18e9-412c-b3d0-2b5e6dcdfed4" (UID: "b75a947d-18e9-412c-b3d0-2b5e6dcdfed4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.286575 4983 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.286635 4983 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-util\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.286645 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2jlb\" (UniqueName: \"kubernetes.io/projected/b75a947d-18e9-412c-b3d0-2b5e6dcdfed4-kube-api-access-j2jlb\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.599563 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" event={"ID":"5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2","Type":"ContainerDied","Data":"9482255574ed9e1b6ab76d94774513b5053c6f636408bbead574f83c68ca82a0"} Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.599599 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.599613 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9482255574ed9e1b6ab76d94774513b5053c6f636408bbead574f83c68ca82a0" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.608827 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" event={"ID":"b75a947d-18e9-412c-b3d0-2b5e6dcdfed4","Type":"ContainerDied","Data":"9017d2caf50666bef96d10c20955944a3b903639685ec60563e8d56d8f6c2c73"} Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.608867 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9017d2caf50666bef96d10c20955944a3b903639685ec60563e8d56d8f6c2c73" Mar 18 00:21:45 crc kubenswrapper[4983]: I0318 00:21:45.609234 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.397650 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx"] Mar 18 00:21:48 crc kubenswrapper[4983]: E0318 00:21:48.398152 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2" containerName="pull" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.398168 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2" containerName="pull" Mar 18 00:21:48 crc kubenswrapper[4983]: E0318 00:21:48.398186 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b75a947d-18e9-412c-b3d0-2b5e6dcdfed4" containerName="pull" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.398195 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="b75a947d-18e9-412c-b3d0-2b5e6dcdfed4" containerName="pull" Mar 18 00:21:48 crc kubenswrapper[4983]: E0318 00:21:48.398209 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b75a947d-18e9-412c-b3d0-2b5e6dcdfed4" containerName="util" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.398217 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="b75a947d-18e9-412c-b3d0-2b5e6dcdfed4" containerName="util" Mar 18 00:21:48 crc kubenswrapper[4983]: E0318 00:21:48.398228 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b75a947d-18e9-412c-b3d0-2b5e6dcdfed4" containerName="extract" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.398236 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="b75a947d-18e9-412c-b3d0-2b5e6dcdfed4" containerName="extract" Mar 18 00:21:48 crc kubenswrapper[4983]: E0318 00:21:48.398249 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2" containerName="util" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.398257 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2" containerName="util" Mar 18 00:21:48 crc kubenswrapper[4983]: E0318 00:21:48.398268 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2" containerName="extract" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.398276 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2" containerName="extract" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.398390 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="b75a947d-18e9-412c-b3d0-2b5e6dcdfed4" containerName="extract" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.398404 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2" containerName="extract" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.399296 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.401998 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.407838 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx"] Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.526927 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2a7d9731-da61-4628-a0f2-2c981292584c-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx\" (UID: \"2a7d9731-da61-4628-a0f2-2c981292584c\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.527006 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnhtl\" (UniqueName: \"kubernetes.io/projected/2a7d9731-da61-4628-a0f2-2c981292584c-kube-api-access-fnhtl\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx\" (UID: \"2a7d9731-da61-4628-a0f2-2c981292584c\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.527036 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2a7d9731-da61-4628-a0f2-2c981292584c-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx\" (UID: \"2a7d9731-da61-4628-a0f2-2c981292584c\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.628276 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2a7d9731-da61-4628-a0f2-2c981292584c-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx\" (UID: \"2a7d9731-da61-4628-a0f2-2c981292584c\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.628324 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnhtl\" (UniqueName: \"kubernetes.io/projected/2a7d9731-da61-4628-a0f2-2c981292584c-kube-api-access-fnhtl\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx\" (UID: \"2a7d9731-da61-4628-a0f2-2c981292584c\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.628347 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2a7d9731-da61-4628-a0f2-2c981292584c-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx\" (UID: \"2a7d9731-da61-4628-a0f2-2c981292584c\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.628781 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2a7d9731-da61-4628-a0f2-2c981292584c-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx\" (UID: \"2a7d9731-da61-4628-a0f2-2c981292584c\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.629016 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2a7d9731-da61-4628-a0f2-2c981292584c-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx\" (UID: \"2a7d9731-da61-4628-a0f2-2c981292584c\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.658000 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnhtl\" (UniqueName: \"kubernetes.io/projected/2a7d9731-da61-4628-a0f2-2c981292584c-kube-api-access-fnhtl\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx\" (UID: \"2a7d9731-da61-4628-a0f2-2c981292584c\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" Mar 18 00:21:48 crc kubenswrapper[4983]: I0318 00:21:48.718611 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" Mar 18 00:21:49 crc kubenswrapper[4983]: I0318 00:21:49.001558 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx"] Mar 18 00:21:49 crc kubenswrapper[4983]: I0318 00:21:49.629929 4983 generic.go:334] "Generic (PLEG): container finished" podID="2a7d9731-da61-4628-a0f2-2c981292584c" containerID="c56dd56a08491ea9a3b5e570400678f576869771f025037b46d23afa02989ce2" exitCode=0 Mar 18 00:21:49 crc kubenswrapper[4983]: I0318 00:21:49.630129 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" event={"ID":"2a7d9731-da61-4628-a0f2-2c981292584c","Type":"ContainerDied","Data":"c56dd56a08491ea9a3b5e570400678f576869771f025037b46d23afa02989ce2"} Mar 18 00:21:49 crc kubenswrapper[4983]: I0318 00:21:49.630276 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" event={"ID":"2a7d9731-da61-4628-a0f2-2c981292584c","Type":"ContainerStarted","Data":"4adb6185d017bc3afbe8a350e37917e214f638016357455ce46a8f964b46b5a4"} Mar 18 00:21:51 crc kubenswrapper[4983]: I0318 00:21:51.684683 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-8ff7d675-d2zl4"] Mar 18 00:21:51 crc kubenswrapper[4983]: I0318 00:21:51.685717 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-d2zl4" Mar 18 00:21:51 crc kubenswrapper[4983]: I0318 00:21:51.687744 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Mar 18 00:21:51 crc kubenswrapper[4983]: I0318 00:21:51.688086 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Mar 18 00:21:51 crc kubenswrapper[4983]: I0318 00:21:51.688106 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-qqkzn" Mar 18 00:21:51 crc kubenswrapper[4983]: I0318 00:21:51.706167 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-8ff7d675-d2zl4"] Mar 18 00:21:51 crc kubenswrapper[4983]: I0318 00:21:51.775193 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq26g\" (UniqueName: \"kubernetes.io/projected/cc11284f-99b8-4e9c-9553-11116529ffec-kube-api-access-xq26g\") pod \"obo-prometheus-operator-8ff7d675-d2zl4\" (UID: \"cc11284f-99b8-4e9c-9553-11116529ffec\") " pod="openshift-operators/obo-prometheus-operator-8ff7d675-d2zl4" Mar 18 00:21:51 crc kubenswrapper[4983]: I0318 00:21:51.876102 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq26g\" (UniqueName: \"kubernetes.io/projected/cc11284f-99b8-4e9c-9553-11116529ffec-kube-api-access-xq26g\") pod \"obo-prometheus-operator-8ff7d675-d2zl4\" (UID: \"cc11284f-99b8-4e9c-9553-11116529ffec\") " pod="openshift-operators/obo-prometheus-operator-8ff7d675-d2zl4" Mar 18 00:21:51 crc kubenswrapper[4983]: I0318 00:21:51.905316 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq26g\" (UniqueName: \"kubernetes.io/projected/cc11284f-99b8-4e9c-9553-11116529ffec-kube-api-access-xq26g\") pod \"obo-prometheus-operator-8ff7d675-d2zl4\" (UID: \"cc11284f-99b8-4e9c-9553-11116529ffec\") " pod="openshift-operators/obo-prometheus-operator-8ff7d675-d2zl4" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.002993 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-8ff7d675-d2zl4" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.050313 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh"] Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.051422 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.053786 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-2mwtp" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.057689 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.064587 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg"] Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.065765 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.066407 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh"] Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.078859 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg"] Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.179656 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7cac8d26-fa2e-4c01-b8ae-16070116f5ea-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg\" (UID: \"7cac8d26-fa2e-4c01-b8ae-16070116f5ea\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.179745 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c1f23aa-597c-4649-bec2-649b188836a3-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh\" (UID: \"1c1f23aa-597c-4649-bec2-649b188836a3\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.180376 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c1f23aa-597c-4649-bec2-649b188836a3-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh\" (UID: \"1c1f23aa-597c-4649-bec2-649b188836a3\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.180499 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7cac8d26-fa2e-4c01-b8ae-16070116f5ea-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg\" (UID: \"7cac8d26-fa2e-4c01-b8ae-16070116f5ea\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.281803 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c1f23aa-597c-4649-bec2-649b188836a3-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh\" (UID: \"1c1f23aa-597c-4649-bec2-649b188836a3\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.281872 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7cac8d26-fa2e-4c01-b8ae-16070116f5ea-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg\" (UID: \"7cac8d26-fa2e-4c01-b8ae-16070116f5ea\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.281951 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7cac8d26-fa2e-4c01-b8ae-16070116f5ea-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg\" (UID: \"7cac8d26-fa2e-4c01-b8ae-16070116f5ea\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.281976 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c1f23aa-597c-4649-bec2-649b188836a3-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh\" (UID: \"1c1f23aa-597c-4649-bec2-649b188836a3\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.285461 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c1f23aa-597c-4649-bec2-649b188836a3-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh\" (UID: \"1c1f23aa-597c-4649-bec2-649b188836a3\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.285861 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7cac8d26-fa2e-4c01-b8ae-16070116f5ea-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg\" (UID: \"7cac8d26-fa2e-4c01-b8ae-16070116f5ea\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.285985 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c1f23aa-597c-4649-bec2-649b188836a3-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh\" (UID: \"1c1f23aa-597c-4649-bec2-649b188836a3\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.295999 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7cac8d26-fa2e-4c01-b8ae-16070116f5ea-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg\" (UID: \"7cac8d26-fa2e-4c01-b8ae-16070116f5ea\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.370221 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.386316 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-6dd7dd855f-9z5qt"] Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.387191 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-6dd7dd855f-9z5qt" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.392002 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-8snzw" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.392222 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.393598 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.407913 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-6dd7dd855f-9z5qt"] Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.484834 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/99cbbe02-ca82-4830-a0a6-00bf36e34545-observability-operator-tls\") pod \"observability-operator-6dd7dd855f-9z5qt\" (UID: \"99cbbe02-ca82-4830-a0a6-00bf36e34545\") " pod="openshift-operators/observability-operator-6dd7dd855f-9z5qt" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.484877 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbwl9\" (UniqueName: \"kubernetes.io/projected/99cbbe02-ca82-4830-a0a6-00bf36e34545-kube-api-access-lbwl9\") pod \"observability-operator-6dd7dd855f-9z5qt\" (UID: \"99cbbe02-ca82-4830-a0a6-00bf36e34545\") " pod="openshift-operators/observability-operator-6dd7dd855f-9z5qt" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.586520 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/99cbbe02-ca82-4830-a0a6-00bf36e34545-observability-operator-tls\") pod \"observability-operator-6dd7dd855f-9z5qt\" (UID: \"99cbbe02-ca82-4830-a0a6-00bf36e34545\") " pod="openshift-operators/observability-operator-6dd7dd855f-9z5qt" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.586611 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbwl9\" (UniqueName: \"kubernetes.io/projected/99cbbe02-ca82-4830-a0a6-00bf36e34545-kube-api-access-lbwl9\") pod \"observability-operator-6dd7dd855f-9z5qt\" (UID: \"99cbbe02-ca82-4830-a0a6-00bf36e34545\") " pod="openshift-operators/observability-operator-6dd7dd855f-9z5qt" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.591477 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/99cbbe02-ca82-4830-a0a6-00bf36e34545-observability-operator-tls\") pod \"observability-operator-6dd7dd855f-9z5qt\" (UID: \"99cbbe02-ca82-4830-a0a6-00bf36e34545\") " pod="openshift-operators/observability-operator-6dd7dd855f-9z5qt" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.618034 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbwl9\" (UniqueName: \"kubernetes.io/projected/99cbbe02-ca82-4830-a0a6-00bf36e34545-kube-api-access-lbwl9\") pod \"observability-operator-6dd7dd855f-9z5qt\" (UID: \"99cbbe02-ca82-4830-a0a6-00bf36e34545\") " pod="openshift-operators/observability-operator-6dd7dd855f-9z5qt" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.704524 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-6dd7dd855f-9z5qt" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.801358 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-749b9786b8-rccft"] Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.802035 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.805062 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-sjb6c" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.806698 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-service-cert" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.818799 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-749b9786b8-rccft"] Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.991671 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/42a01360-4470-43f7-a2a5-e3262ba6ab1c-openshift-service-ca\") pod \"perses-operator-749b9786b8-rccft\" (UID: \"42a01360-4470-43f7-a2a5-e3262ba6ab1c\") " pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.992024 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d864j\" (UniqueName: \"kubernetes.io/projected/42a01360-4470-43f7-a2a5-e3262ba6ab1c-kube-api-access-d864j\") pod \"perses-operator-749b9786b8-rccft\" (UID: \"42a01360-4470-43f7-a2a5-e3262ba6ab1c\") " pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.992085 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/42a01360-4470-43f7-a2a5-e3262ba6ab1c-apiservice-cert\") pod \"perses-operator-749b9786b8-rccft\" (UID: \"42a01360-4470-43f7-a2a5-e3262ba6ab1c\") " pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:52 crc kubenswrapper[4983]: I0318 00:21:52.992121 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/42a01360-4470-43f7-a2a5-e3262ba6ab1c-webhook-cert\") pod \"perses-operator-749b9786b8-rccft\" (UID: \"42a01360-4470-43f7-a2a5-e3262ba6ab1c\") " pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:53 crc kubenswrapper[4983]: I0318 00:21:53.093871 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/42a01360-4470-43f7-a2a5-e3262ba6ab1c-openshift-service-ca\") pod \"perses-operator-749b9786b8-rccft\" (UID: \"42a01360-4470-43f7-a2a5-e3262ba6ab1c\") " pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:53 crc kubenswrapper[4983]: I0318 00:21:53.093919 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d864j\" (UniqueName: \"kubernetes.io/projected/42a01360-4470-43f7-a2a5-e3262ba6ab1c-kube-api-access-d864j\") pod \"perses-operator-749b9786b8-rccft\" (UID: \"42a01360-4470-43f7-a2a5-e3262ba6ab1c\") " pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:53 crc kubenswrapper[4983]: I0318 00:21:53.093969 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/42a01360-4470-43f7-a2a5-e3262ba6ab1c-apiservice-cert\") pod \"perses-operator-749b9786b8-rccft\" (UID: \"42a01360-4470-43f7-a2a5-e3262ba6ab1c\") " pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:53 crc kubenswrapper[4983]: I0318 00:21:53.093996 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/42a01360-4470-43f7-a2a5-e3262ba6ab1c-webhook-cert\") pod \"perses-operator-749b9786b8-rccft\" (UID: \"42a01360-4470-43f7-a2a5-e3262ba6ab1c\") " pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:53 crc kubenswrapper[4983]: I0318 00:21:53.094780 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/42a01360-4470-43f7-a2a5-e3262ba6ab1c-openshift-service-ca\") pod \"perses-operator-749b9786b8-rccft\" (UID: \"42a01360-4470-43f7-a2a5-e3262ba6ab1c\") " pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:53 crc kubenswrapper[4983]: I0318 00:21:53.099750 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/42a01360-4470-43f7-a2a5-e3262ba6ab1c-apiservice-cert\") pod \"perses-operator-749b9786b8-rccft\" (UID: \"42a01360-4470-43f7-a2a5-e3262ba6ab1c\") " pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:53 crc kubenswrapper[4983]: I0318 00:21:53.099950 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/42a01360-4470-43f7-a2a5-e3262ba6ab1c-webhook-cert\") pod \"perses-operator-749b9786b8-rccft\" (UID: \"42a01360-4470-43f7-a2a5-e3262ba6ab1c\") " pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:53 crc kubenswrapper[4983]: I0318 00:21:53.109711 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d864j\" (UniqueName: \"kubernetes.io/projected/42a01360-4470-43f7-a2a5-e3262ba6ab1c-kube-api-access-d864j\") pod \"perses-operator-749b9786b8-rccft\" (UID: \"42a01360-4470-43f7-a2a5-e3262ba6ab1c\") " pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:53 crc kubenswrapper[4983]: I0318 00:21:53.118580 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:21:54 crc kubenswrapper[4983]: I0318 00:21:54.388474 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-6dd7dd855f-9z5qt"] Mar 18 00:21:54 crc kubenswrapper[4983]: W0318 00:21:54.413877 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99cbbe02_ca82_4830_a0a6_00bf36e34545.slice/crio-223a17fa728ae968bcc8c92d2f02dde686e4b4003faa8617ab07f758319f615b WatchSource:0}: Error finding container 223a17fa728ae968bcc8c92d2f02dde686e4b4003faa8617ab07f758319f615b: Status 404 returned error can't find the container with id 223a17fa728ae968bcc8c92d2f02dde686e4b4003faa8617ab07f758319f615b Mar 18 00:21:54 crc kubenswrapper[4983]: I0318 00:21:54.664533 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-6dd7dd855f-9z5qt" event={"ID":"99cbbe02-ca82-4830-a0a6-00bf36e34545","Type":"ContainerStarted","Data":"223a17fa728ae968bcc8c92d2f02dde686e4b4003faa8617ab07f758319f615b"} Mar 18 00:21:54 crc kubenswrapper[4983]: I0318 00:21:54.666511 4983 generic.go:334] "Generic (PLEG): container finished" podID="2a7d9731-da61-4628-a0f2-2c981292584c" containerID="3d24040e364e94f15edba2dde1152f569679cc04212b84ee4b8eb8b16bd26755" exitCode=0 Mar 18 00:21:54 crc kubenswrapper[4983]: I0318 00:21:54.666546 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" event={"ID":"2a7d9731-da61-4628-a0f2-2c981292584c","Type":"ContainerDied","Data":"3d24040e364e94f15edba2dde1152f569679cc04212b84ee4b8eb8b16bd26755"} Mar 18 00:21:54 crc kubenswrapper[4983]: I0318 00:21:54.680079 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg"] Mar 18 00:21:54 crc kubenswrapper[4983]: I0318 00:21:54.759757 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-749b9786b8-rccft"] Mar 18 00:21:54 crc kubenswrapper[4983]: W0318 00:21:54.774400 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42a01360_4470_43f7_a2a5_e3262ba6ab1c.slice/crio-d7bfed8d1e33171dbe5ac20b9d537305a7915d18c4929a6e76c0ec5ac0c898a2 WatchSource:0}: Error finding container d7bfed8d1e33171dbe5ac20b9d537305a7915d18c4929a6e76c0ec5ac0c898a2: Status 404 returned error can't find the container with id d7bfed8d1e33171dbe5ac20b9d537305a7915d18c4929a6e76c0ec5ac0c898a2 Mar 18 00:21:54 crc kubenswrapper[4983]: I0318 00:21:54.785770 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-8ff7d675-d2zl4"] Mar 18 00:21:54 crc kubenswrapper[4983]: W0318 00:21:54.798910 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc11284f_99b8_4e9c_9553_11116529ffec.slice/crio-fd00f5af4f2dcfe02b153a17924370e4b4aee5bdaaaf7d714368a8cb36956758 WatchSource:0}: Error finding container fd00f5af4f2dcfe02b153a17924370e4b4aee5bdaaaf7d714368a8cb36956758: Status 404 returned error can't find the container with id fd00f5af4f2dcfe02b153a17924370e4b4aee5bdaaaf7d714368a8cb36956758 Mar 18 00:21:54 crc kubenswrapper[4983]: I0318 00:21:54.828957 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh"] Mar 18 00:21:55 crc kubenswrapper[4983]: I0318 00:21:55.675473 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-749b9786b8-rccft" event={"ID":"42a01360-4470-43f7-a2a5-e3262ba6ab1c","Type":"ContainerStarted","Data":"d7bfed8d1e33171dbe5ac20b9d537305a7915d18c4929a6e76c0ec5ac0c898a2"} Mar 18 00:21:55 crc kubenswrapper[4983]: I0318 00:21:55.702560 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-8ff7d675-d2zl4" event={"ID":"cc11284f-99b8-4e9c-9553-11116529ffec","Type":"ContainerStarted","Data":"fd00f5af4f2dcfe02b153a17924370e4b4aee5bdaaaf7d714368a8cb36956758"} Mar 18 00:21:55 crc kubenswrapper[4983]: I0318 00:21:55.709159 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh" event={"ID":"1c1f23aa-597c-4649-bec2-649b188836a3","Type":"ContainerStarted","Data":"72be77bcbe5d747e9e40dfbb8d4dc95ede1d4e69d76cd0862cecda380551b2bf"} Mar 18 00:21:55 crc kubenswrapper[4983]: I0318 00:21:55.735915 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg" event={"ID":"7cac8d26-fa2e-4c01-b8ae-16070116f5ea","Type":"ContainerStarted","Data":"a3c7a66625af2b24a634a67fd406306e28594f648676399de17df4181c50422a"} Mar 18 00:21:55 crc kubenswrapper[4983]: I0318 00:21:55.758618 4983 generic.go:334] "Generic (PLEG): container finished" podID="2a7d9731-da61-4628-a0f2-2c981292584c" containerID="af18bcd5d9a85c1260960926ec7ae0de273fed242c8b3c3a0267c297f52b2137" exitCode=0 Mar 18 00:21:55 crc kubenswrapper[4983]: I0318 00:21:55.758663 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" event={"ID":"2a7d9731-da61-4628-a0f2-2c981292584c","Type":"ContainerDied","Data":"af18bcd5d9a85c1260960926ec7ae0de273fed242c8b3c3a0267c297f52b2137"} Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.236254 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.273484 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2a7d9731-da61-4628-a0f2-2c981292584c-bundle\") pod \"2a7d9731-da61-4628-a0f2-2c981292584c\" (UID: \"2a7d9731-da61-4628-a0f2-2c981292584c\") " Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.273614 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnhtl\" (UniqueName: \"kubernetes.io/projected/2a7d9731-da61-4628-a0f2-2c981292584c-kube-api-access-fnhtl\") pod \"2a7d9731-da61-4628-a0f2-2c981292584c\" (UID: \"2a7d9731-da61-4628-a0f2-2c981292584c\") " Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.273730 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2a7d9731-da61-4628-a0f2-2c981292584c-util\") pod \"2a7d9731-da61-4628-a0f2-2c981292584c\" (UID: \"2a7d9731-da61-4628-a0f2-2c981292584c\") " Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.276153 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a7d9731-da61-4628-a0f2-2c981292584c-bundle" (OuterVolumeSpecName: "bundle") pod "2a7d9731-da61-4628-a0f2-2c981292584c" (UID: "2a7d9731-da61-4628-a0f2-2c981292584c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.285127 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a7d9731-da61-4628-a0f2-2c981292584c-kube-api-access-fnhtl" (OuterVolumeSpecName: "kube-api-access-fnhtl") pod "2a7d9731-da61-4628-a0f2-2c981292584c" (UID: "2a7d9731-da61-4628-a0f2-2c981292584c"). InnerVolumeSpecName "kube-api-access-fnhtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.297918 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a7d9731-da61-4628-a0f2-2c981292584c-util" (OuterVolumeSpecName: "util") pod "2a7d9731-da61-4628-a0f2-2c981292584c" (UID: "2a7d9731-da61-4628-a0f2-2c981292584c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.374846 4983 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2a7d9731-da61-4628-a0f2-2c981292584c-util\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.374885 4983 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2a7d9731-da61-4628-a0f2-2c981292584c-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.375196 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnhtl\" (UniqueName: \"kubernetes.io/projected/2a7d9731-da61-4628-a0f2-2c981292584c-kube-api-access-fnhtl\") on node \"crc\" DevicePath \"\"" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.642491 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-6c5b5c6864-fd5fk"] Mar 18 00:21:57 crc kubenswrapper[4983]: E0318 00:21:57.642719 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a7d9731-da61-4628-a0f2-2c981292584c" containerName="pull" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.642731 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a7d9731-da61-4628-a0f2-2c981292584c" containerName="pull" Mar 18 00:21:57 crc kubenswrapper[4983]: E0318 00:21:57.642746 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a7d9731-da61-4628-a0f2-2c981292584c" containerName="util" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.642751 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a7d9731-da61-4628-a0f2-2c981292584c" containerName="util" Mar 18 00:21:57 crc kubenswrapper[4983]: E0318 00:21:57.642762 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a7d9731-da61-4628-a0f2-2c981292584c" containerName="extract" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.642769 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a7d9731-da61-4628-a0f2-2c981292584c" containerName="extract" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.642854 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a7d9731-da61-4628-a0f2-2c981292584c" containerName="extract" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.643334 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.645086 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-mwbcj" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.645470 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.645682 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.646023 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.654761 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-6c5b5c6864-fd5fk"] Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.780246 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nhrf\" (UniqueName: \"kubernetes.io/projected/b18e78af-5735-4edd-a0c7-6305ec39f81c-kube-api-access-6nhrf\") pod \"elastic-operator-6c5b5c6864-fd5fk\" (UID: \"b18e78af-5735-4edd-a0c7-6305ec39f81c\") " pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.780295 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b18e78af-5735-4edd-a0c7-6305ec39f81c-apiservice-cert\") pod \"elastic-operator-6c5b5c6864-fd5fk\" (UID: \"b18e78af-5735-4edd-a0c7-6305ec39f81c\") " pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.780319 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b18e78af-5735-4edd-a0c7-6305ec39f81c-webhook-cert\") pod \"elastic-operator-6c5b5c6864-fd5fk\" (UID: \"b18e78af-5735-4edd-a0c7-6305ec39f81c\") " pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.791148 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" event={"ID":"2a7d9731-da61-4628-a0f2-2c981292584c","Type":"ContainerDied","Data":"4adb6185d017bc3afbe8a350e37917e214f638016357455ce46a8f964b46b5a4"} Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.791233 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4adb6185d017bc3afbe8a350e37917e214f638016357455ce46a8f964b46b5a4" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.791344 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.883812 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b18e78af-5735-4edd-a0c7-6305ec39f81c-apiservice-cert\") pod \"elastic-operator-6c5b5c6864-fd5fk\" (UID: \"b18e78af-5735-4edd-a0c7-6305ec39f81c\") " pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.883857 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b18e78af-5735-4edd-a0c7-6305ec39f81c-webhook-cert\") pod \"elastic-operator-6c5b5c6864-fd5fk\" (UID: \"b18e78af-5735-4edd-a0c7-6305ec39f81c\") " pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.883922 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nhrf\" (UniqueName: \"kubernetes.io/projected/b18e78af-5735-4edd-a0c7-6305ec39f81c-kube-api-access-6nhrf\") pod \"elastic-operator-6c5b5c6864-fd5fk\" (UID: \"b18e78af-5735-4edd-a0c7-6305ec39f81c\") " pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.890810 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b18e78af-5735-4edd-a0c7-6305ec39f81c-webhook-cert\") pod \"elastic-operator-6c5b5c6864-fd5fk\" (UID: \"b18e78af-5735-4edd-a0c7-6305ec39f81c\") " pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.891310 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b18e78af-5735-4edd-a0c7-6305ec39f81c-apiservice-cert\") pod \"elastic-operator-6c5b5c6864-fd5fk\" (UID: \"b18e78af-5735-4edd-a0c7-6305ec39f81c\") " pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.903306 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nhrf\" (UniqueName: \"kubernetes.io/projected/b18e78af-5735-4edd-a0c7-6305ec39f81c-kube-api-access-6nhrf\") pod \"elastic-operator-6c5b5c6864-fd5fk\" (UID: \"b18e78af-5735-4edd-a0c7-6305ec39f81c\") " pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" Mar 18 00:21:57 crc kubenswrapper[4983]: I0318 00:21:57.972778 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" Mar 18 00:21:58 crc kubenswrapper[4983]: I0318 00:21:58.537780 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-6c5b5c6864-fd5fk"] Mar 18 00:21:58 crc kubenswrapper[4983]: W0318 00:21:58.583033 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb18e78af_5735_4edd_a0c7_6305ec39f81c.slice/crio-9d04cfe3cf464581a816d6f9a2dd2982143b10b6e828391683b18ca9d88417a1 WatchSource:0}: Error finding container 9d04cfe3cf464581a816d6f9a2dd2982143b10b6e828391683b18ca9d88417a1: Status 404 returned error can't find the container with id 9d04cfe3cf464581a816d6f9a2dd2982143b10b6e828391683b18ca9d88417a1 Mar 18 00:21:58 crc kubenswrapper[4983]: I0318 00:21:58.800767 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" event={"ID":"b18e78af-5735-4edd-a0c7-6305ec39f81c","Type":"ContainerStarted","Data":"9d04cfe3cf464581a816d6f9a2dd2982143b10b6e828391683b18ca9d88417a1"} Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.150060 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563222-57gsb"] Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.151426 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563222-57gsb" Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.155575 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.155635 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.155726 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.157052 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563222-57gsb"] Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.321083 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfdh9\" (UniqueName: \"kubernetes.io/projected/c99a06e4-c46d-4036-8307-9eabfed2b5b0-kube-api-access-vfdh9\") pod \"auto-csr-approver-29563222-57gsb\" (UID: \"c99a06e4-c46d-4036-8307-9eabfed2b5b0\") " pod="openshift-infra/auto-csr-approver-29563222-57gsb" Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.422357 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfdh9\" (UniqueName: \"kubernetes.io/projected/c99a06e4-c46d-4036-8307-9eabfed2b5b0-kube-api-access-vfdh9\") pod \"auto-csr-approver-29563222-57gsb\" (UID: \"c99a06e4-c46d-4036-8307-9eabfed2b5b0\") " pod="openshift-infra/auto-csr-approver-29563222-57gsb" Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.447003 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfdh9\" (UniqueName: \"kubernetes.io/projected/c99a06e4-c46d-4036-8307-9eabfed2b5b0-kube-api-access-vfdh9\") pod \"auto-csr-approver-29563222-57gsb\" (UID: \"c99a06e4-c46d-4036-8307-9eabfed2b5b0\") " pod="openshift-infra/auto-csr-approver-29563222-57gsb" Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.484162 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563222-57gsb" Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.654391 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-l49nf"] Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.655499 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-l49nf" Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.660371 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-wnhjl" Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.666070 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-l49nf"] Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.828665 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkjlb\" (UniqueName: \"kubernetes.io/projected/2fec475a-0660-46b4-b502-a752451f2696-kube-api-access-vkjlb\") pod \"interconnect-operator-5bb49f789d-l49nf\" (UID: \"2fec475a-0660-46b4-b502-a752451f2696\") " pod="service-telemetry/interconnect-operator-5bb49f789d-l49nf" Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.863903 4983 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.933647 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkjlb\" (UniqueName: \"kubernetes.io/projected/2fec475a-0660-46b4-b502-a752451f2696-kube-api-access-vkjlb\") pod \"interconnect-operator-5bb49f789d-l49nf\" (UID: \"2fec475a-0660-46b4-b502-a752451f2696\") " pod="service-telemetry/interconnect-operator-5bb49f789d-l49nf" Mar 18 00:22:00 crc kubenswrapper[4983]: I0318 00:22:00.968548 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkjlb\" (UniqueName: \"kubernetes.io/projected/2fec475a-0660-46b4-b502-a752451f2696-kube-api-access-vkjlb\") pod \"interconnect-operator-5bb49f789d-l49nf\" (UID: \"2fec475a-0660-46b4-b502-a752451f2696\") " pod="service-telemetry/interconnect-operator-5bb49f789d-l49nf" Mar 18 00:22:01 crc kubenswrapper[4983]: I0318 00:22:01.026389 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-l49nf" Mar 18 00:22:08 crc kubenswrapper[4983]: E0318 00:22:08.959920 4983 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:f78b160ba3b815f53d6a72425f3f3a9d7946795177bd68c7c614fa84f97be630" Mar 18 00:22:08 crc kubenswrapper[4983]: E0318 00:22:08.961755 4983 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:f78b160ba3b815f53d6a72425f3f3a9d7946795177bd68c7c614fa84f97be630,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.4.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d864j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-749b9786b8-rccft_openshift-operators(42a01360-4470-43f7-a2a5-e3262ba6ab1c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 18 00:22:08 crc kubenswrapper[4983]: E0318 00:22:08.963050 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-749b9786b8-rccft" podUID="42a01360-4470-43f7-a2a5-e3262ba6ab1c" Mar 18 00:22:09 crc kubenswrapper[4983]: E0318 00:22:09.920288 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:f78b160ba3b815f53d6a72425f3f3a9d7946795177bd68c7c614fa84f97be630\\\"\"" pod="openshift-operators/perses-operator-749b9786b8-rccft" podUID="42a01360-4470-43f7-a2a5-e3262ba6ab1c" Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.635830 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563222-57gsb"] Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.674064 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-l49nf"] Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.920545 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg" event={"ID":"7cac8d26-fa2e-4c01-b8ae-16070116f5ea","Type":"ContainerStarted","Data":"dd64d6ab44caa3f903015e30a271baa663a711aedda5ce2f2357119a7210d472"} Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.922905 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" event={"ID":"b18e78af-5735-4edd-a0c7-6305ec39f81c","Type":"ContainerStarted","Data":"ddaecf556513054c8e19d35bd7cef468d334e221418308c64b64f78c62d689ef"} Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.924148 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563222-57gsb" event={"ID":"c99a06e4-c46d-4036-8307-9eabfed2b5b0","Type":"ContainerStarted","Data":"7b1a6c9cd02ba0e86f7ed9823d1e0f37dc1e2cfb9ca01e8e2fff7984bf3b9dd1"} Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.925340 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-l49nf" event={"ID":"2fec475a-0660-46b4-b502-a752451f2696","Type":"ContainerStarted","Data":"9a54821e8e79a26e29018a199cf9b37df625863131a3a21fc9919c33b4a62340"} Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.927119 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-6dd7dd855f-9z5qt" event={"ID":"99cbbe02-ca82-4830-a0a6-00bf36e34545","Type":"ContainerStarted","Data":"bcb0b782d6c3fb955b610baf881187f4ce6bb8e351187af667fc3439257307ac"} Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.927350 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-6dd7dd855f-9z5qt" Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.928668 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-8ff7d675-d2zl4" event={"ID":"cc11284f-99b8-4e9c-9553-11116529ffec","Type":"ContainerStarted","Data":"c33700908ffb8085c1dd817e40fccc69dd86983ca5b4b6658365ccebf3fbb2b8"} Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.930684 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh" event={"ID":"1c1f23aa-597c-4649-bec2-649b188836a3","Type":"ContainerStarted","Data":"f594b562c455da2dd53bee21f3b27f71302dfda8c72b7f1b954a65f9ae754ea5"} Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.938136 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-6dd7dd855f-9z5qt" Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.941518 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg" podStartSLOduration=3.2979547240000002 podStartE2EDuration="18.941497491s" podCreationTimestamp="2026-03-18 00:21:52 +0000 UTC" firstStartedPulling="2026-03-18 00:21:54.703520821 +0000 UTC m=+832.600895431" lastFinishedPulling="2026-03-18 00:22:10.347063588 +0000 UTC m=+848.244438198" observedRunningTime="2026-03-18 00:22:10.937917289 +0000 UTC m=+848.835291919" watchObservedRunningTime="2026-03-18 00:22:10.941497491 +0000 UTC m=+848.838872101" Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.972904 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-6dd7dd855f-9z5qt" podStartSLOduration=3.050669018 podStartE2EDuration="18.972886871s" podCreationTimestamp="2026-03-18 00:21:52 +0000 UTC" firstStartedPulling="2026-03-18 00:21:54.424087534 +0000 UTC m=+832.321462144" lastFinishedPulling="2026-03-18 00:22:10.346305377 +0000 UTC m=+848.243679997" observedRunningTime="2026-03-18 00:22:10.969216447 +0000 UTC m=+848.866591067" watchObservedRunningTime="2026-03-18 00:22:10.972886871 +0000 UTC m=+848.870261481" Mar 18 00:22:10 crc kubenswrapper[4983]: I0318 00:22:10.994183 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh" podStartSLOduration=4.153637656 podStartE2EDuration="18.994163905s" podCreationTimestamp="2026-03-18 00:21:52 +0000 UTC" firstStartedPulling="2026-03-18 00:21:54.851770746 +0000 UTC m=+832.749145356" lastFinishedPulling="2026-03-18 00:22:09.692296995 +0000 UTC m=+847.589671605" observedRunningTime="2026-03-18 00:22:10.99364275 +0000 UTC m=+848.891017380" watchObservedRunningTime="2026-03-18 00:22:10.994163905 +0000 UTC m=+848.891538515" Mar 18 00:22:11 crc kubenswrapper[4983]: I0318 00:22:11.034002 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-6c5b5c6864-fd5fk" podStartSLOduration=2.152569286 podStartE2EDuration="14.033981264s" podCreationTimestamp="2026-03-18 00:21:57 +0000 UTC" firstStartedPulling="2026-03-18 00:21:58.612530287 +0000 UTC m=+836.509904897" lastFinishedPulling="2026-03-18 00:22:10.493942265 +0000 UTC m=+848.391316875" observedRunningTime="2026-03-18 00:22:11.03102003 +0000 UTC m=+848.928394650" watchObservedRunningTime="2026-03-18 00:22:11.033981264 +0000 UTC m=+848.931355874" Mar 18 00:22:11 crc kubenswrapper[4983]: I0318 00:22:11.074781 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-8ff7d675-d2zl4" podStartSLOduration=5.185365507 podStartE2EDuration="20.074762821s" podCreationTimestamp="2026-03-18 00:21:51 +0000 UTC" firstStartedPulling="2026-03-18 00:21:54.8021848 +0000 UTC m=+832.699559410" lastFinishedPulling="2026-03-18 00:22:09.691582114 +0000 UTC m=+847.588956724" observedRunningTime="2026-03-18 00:22:11.07048468 +0000 UTC m=+848.967859300" watchObservedRunningTime="2026-03-18 00:22:11.074762821 +0000 UTC m=+848.972137431" Mar 18 00:22:11 crc kubenswrapper[4983]: I0318 00:22:11.978776 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 18 00:22:11 crc kubenswrapper[4983]: I0318 00:22:11.981415 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:11 crc kubenswrapper[4983]: I0318 00:22:11.989757 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Mar 18 00:22:11 crc kubenswrapper[4983]: I0318 00:22:11.990253 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Mar 18 00:22:11 crc kubenswrapper[4983]: I0318 00:22:11.990528 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.004383 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-wl7h7" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.004888 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.006162 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.006491 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.006710 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.008634 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.025717 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117066 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117124 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117152 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117185 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/682dc305-300d-4702-a640-b422a2185661-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117204 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117225 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117243 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/682dc305-300d-4702-a640-b422a2185661-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117402 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117445 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117487 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117537 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/682dc305-300d-4702-a640-b422a2185661-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117556 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117588 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117608 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.117638 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.220074 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.220580 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.220643 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.220673 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.220698 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.220720 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.220743 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/682dc305-300d-4702-a640-b422a2185661-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.220763 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.220780 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.220788 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.221904 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.222217 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.222300 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.220801 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/682dc305-300d-4702-a640-b422a2185661-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.222342 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.222365 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.222395 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.222427 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/682dc305-300d-4702-a640-b422a2185661-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.222429 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/682dc305-300d-4702-a640-b422a2185661-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.222445 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.224816 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/682dc305-300d-4702-a640-b422a2185661-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.233875 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.233997 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.234593 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.234683 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.236502 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/682dc305-300d-4702-a640-b422a2185661-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.236775 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.238453 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/682dc305-300d-4702-a640-b422a2185661-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.252304 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.269555 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/682dc305-300d-4702-a640-b422a2185661-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"682dc305-300d-4702-a640-b422a2185661\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.317859 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.748601 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 18 00:22:12 crc kubenswrapper[4983]: W0318 00:22:12.804776 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod682dc305_300d_4702_a640_b422a2185661.slice/crio-164046f05431397f11a8d87923b529a176e13baa07b4696d2591b2ecf83b323b WatchSource:0}: Error finding container 164046f05431397f11a8d87923b529a176e13baa07b4696d2591b2ecf83b323b: Status 404 returned error can't find the container with id 164046f05431397f11a8d87923b529a176e13baa07b4696d2591b2ecf83b323b Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.961401 4983 generic.go:334] "Generic (PLEG): container finished" podID="c99a06e4-c46d-4036-8307-9eabfed2b5b0" containerID="6fb5448b2b229082f07dc8f36cd95fb67a21f88f4883fbe632c3dd0e769ecb02" exitCode=0 Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.961484 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563222-57gsb" event={"ID":"c99a06e4-c46d-4036-8307-9eabfed2b5b0","Type":"ContainerDied","Data":"6fb5448b2b229082f07dc8f36cd95fb67a21f88f4883fbe632c3dd0e769ecb02"} Mar 18 00:22:12 crc kubenswrapper[4983]: I0318 00:22:12.963361 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"682dc305-300d-4702-a640-b422a2185661","Type":"ContainerStarted","Data":"164046f05431397f11a8d87923b529a176e13baa07b4696d2591b2ecf83b323b"} Mar 18 00:22:14 crc kubenswrapper[4983]: I0318 00:22:14.287782 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563222-57gsb" Mar 18 00:22:14 crc kubenswrapper[4983]: I0318 00:22:14.368220 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfdh9\" (UniqueName: \"kubernetes.io/projected/c99a06e4-c46d-4036-8307-9eabfed2b5b0-kube-api-access-vfdh9\") pod \"c99a06e4-c46d-4036-8307-9eabfed2b5b0\" (UID: \"c99a06e4-c46d-4036-8307-9eabfed2b5b0\") " Mar 18 00:22:14 crc kubenswrapper[4983]: I0318 00:22:14.377407 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c99a06e4-c46d-4036-8307-9eabfed2b5b0-kube-api-access-vfdh9" (OuterVolumeSpecName: "kube-api-access-vfdh9") pod "c99a06e4-c46d-4036-8307-9eabfed2b5b0" (UID: "c99a06e4-c46d-4036-8307-9eabfed2b5b0"). InnerVolumeSpecName "kube-api-access-vfdh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:22:14 crc kubenswrapper[4983]: I0318 00:22:14.471822 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfdh9\" (UniqueName: \"kubernetes.io/projected/c99a06e4-c46d-4036-8307-9eabfed2b5b0-kube-api-access-vfdh9\") on node \"crc\" DevicePath \"\"" Mar 18 00:22:15 crc kubenswrapper[4983]: I0318 00:22:15.022556 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563222-57gsb" event={"ID":"c99a06e4-c46d-4036-8307-9eabfed2b5b0","Type":"ContainerDied","Data":"7b1a6c9cd02ba0e86f7ed9823d1e0f37dc1e2cfb9ca01e8e2fff7984bf3b9dd1"} Mar 18 00:22:15 crc kubenswrapper[4983]: I0318 00:22:15.022628 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b1a6c9cd02ba0e86f7ed9823d1e0f37dc1e2cfb9ca01e8e2fff7984bf3b9dd1" Mar 18 00:22:15 crc kubenswrapper[4983]: I0318 00:22:15.022737 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563222-57gsb" Mar 18 00:22:15 crc kubenswrapper[4983]: I0318 00:22:15.347156 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563216-jj8fm"] Mar 18 00:22:15 crc kubenswrapper[4983]: I0318 00:22:15.350496 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563216-jj8fm"] Mar 18 00:22:16 crc kubenswrapper[4983]: I0318 00:22:16.859581 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf" path="/var/lib/kubelet/pods/a27c2d9e-bdb6-4502-b0ba-24bd0bfeefdf/volumes" Mar 18 00:22:16 crc kubenswrapper[4983]: I0318 00:22:16.860587 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9"] Mar 18 00:22:16 crc kubenswrapper[4983]: E0318 00:22:16.860922 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c99a06e4-c46d-4036-8307-9eabfed2b5b0" containerName="oc" Mar 18 00:22:16 crc kubenswrapper[4983]: I0318 00:22:16.860962 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="c99a06e4-c46d-4036-8307-9eabfed2b5b0" containerName="oc" Mar 18 00:22:16 crc kubenswrapper[4983]: I0318 00:22:16.861145 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="c99a06e4-c46d-4036-8307-9eabfed2b5b0" containerName="oc" Mar 18 00:22:16 crc kubenswrapper[4983]: I0318 00:22:16.861921 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9" Mar 18 00:22:16 crc kubenswrapper[4983]: I0318 00:22:16.867793 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Mar 18 00:22:16 crc kubenswrapper[4983]: I0318 00:22:16.868609 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Mar 18 00:22:16 crc kubenswrapper[4983]: I0318 00:22:16.868789 4983 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-fdk5s" Mar 18 00:22:16 crc kubenswrapper[4983]: I0318 00:22:16.893343 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9"] Mar 18 00:22:16 crc kubenswrapper[4983]: I0318 00:22:16.913633 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9e82f3b2-4502-4ea4-9fc7-55510b8e75a4-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-fjft9\" (UID: \"9e82f3b2-4502-4ea4-9fc7-55510b8e75a4\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9" Mar 18 00:22:16 crc kubenswrapper[4983]: I0318 00:22:16.913766 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hjn7\" (UniqueName: \"kubernetes.io/projected/9e82f3b2-4502-4ea4-9fc7-55510b8e75a4-kube-api-access-2hjn7\") pod \"cert-manager-operator-controller-manager-5586865c96-fjft9\" (UID: \"9e82f3b2-4502-4ea4-9fc7-55510b8e75a4\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9" Mar 18 00:22:17 crc kubenswrapper[4983]: I0318 00:22:17.015826 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9e82f3b2-4502-4ea4-9fc7-55510b8e75a4-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-fjft9\" (UID: \"9e82f3b2-4502-4ea4-9fc7-55510b8e75a4\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9" Mar 18 00:22:17 crc kubenswrapper[4983]: I0318 00:22:17.015911 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hjn7\" (UniqueName: \"kubernetes.io/projected/9e82f3b2-4502-4ea4-9fc7-55510b8e75a4-kube-api-access-2hjn7\") pod \"cert-manager-operator-controller-manager-5586865c96-fjft9\" (UID: \"9e82f3b2-4502-4ea4-9fc7-55510b8e75a4\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9" Mar 18 00:22:17 crc kubenswrapper[4983]: I0318 00:22:17.016436 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9e82f3b2-4502-4ea4-9fc7-55510b8e75a4-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-fjft9\" (UID: \"9e82f3b2-4502-4ea4-9fc7-55510b8e75a4\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9" Mar 18 00:22:17 crc kubenswrapper[4983]: I0318 00:22:17.041956 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hjn7\" (UniqueName: \"kubernetes.io/projected/9e82f3b2-4502-4ea4-9fc7-55510b8e75a4-kube-api-access-2hjn7\") pod \"cert-manager-operator-controller-manager-5586865c96-fjft9\" (UID: \"9e82f3b2-4502-4ea4-9fc7-55510b8e75a4\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9" Mar 18 00:22:17 crc kubenswrapper[4983]: I0318 00:22:17.177148 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9" Mar 18 00:22:29 crc kubenswrapper[4983]: I0318 00:22:29.040160 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9"] Mar 18 00:22:29 crc kubenswrapper[4983]: W0318 00:22:29.050806 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e82f3b2_4502_4ea4_9fc7_55510b8e75a4.slice/crio-c710b5e316d996bc6b2f4dce468fc800975542bffb7b5e87aed03b30b8b33035 WatchSource:0}: Error finding container c710b5e316d996bc6b2f4dce468fc800975542bffb7b5e87aed03b30b8b33035: Status 404 returned error can't find the container with id c710b5e316d996bc6b2f4dce468fc800975542bffb7b5e87aed03b30b8b33035 Mar 18 00:22:29 crc kubenswrapper[4983]: I0318 00:22:29.141361 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9" event={"ID":"9e82f3b2-4502-4ea4-9fc7-55510b8e75a4","Type":"ContainerStarted","Data":"c710b5e316d996bc6b2f4dce468fc800975542bffb7b5e87aed03b30b8b33035"} Mar 18 00:22:30 crc kubenswrapper[4983]: I0318 00:22:30.158482 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-l49nf" event={"ID":"2fec475a-0660-46b4-b502-a752451f2696","Type":"ContainerStarted","Data":"9c2df8f8b4c4df4c9ec794aaa122e4ecd42035a4b5f46ec4ad75fee3ee29c64a"} Mar 18 00:22:30 crc kubenswrapper[4983]: I0318 00:22:30.160898 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-749b9786b8-rccft" event={"ID":"42a01360-4470-43f7-a2a5-e3262ba6ab1c","Type":"ContainerStarted","Data":"cbed571ee97c2626a2b9556374409b290fec7284073e0f13df0740b3a0d453fa"} Mar 18 00:22:30 crc kubenswrapper[4983]: I0318 00:22:30.161760 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:22:30 crc kubenswrapper[4983]: I0318 00:22:30.163814 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"682dc305-300d-4702-a640-b422a2185661","Type":"ContainerStarted","Data":"7cc6f1458252ccd85a9b29d27114bda72dc5482bba95dac2358c7859f23b00c5"} Mar 18 00:22:30 crc kubenswrapper[4983]: I0318 00:22:30.178954 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-l49nf" podStartSLOduration=12.183153222 podStartE2EDuration="30.178915454s" podCreationTimestamp="2026-03-18 00:22:00 +0000 UTC" firstStartedPulling="2026-03-18 00:22:10.703539441 +0000 UTC m=+848.600914051" lastFinishedPulling="2026-03-18 00:22:28.699301673 +0000 UTC m=+866.596676283" observedRunningTime="2026-03-18 00:22:30.177543145 +0000 UTC m=+868.074917755" watchObservedRunningTime="2026-03-18 00:22:30.178915454 +0000 UTC m=+868.076290064" Mar 18 00:22:30 crc kubenswrapper[4983]: I0318 00:22:30.209186 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-749b9786b8-rccft" podStartSLOduration=4.221250984 podStartE2EDuration="38.209159192s" podCreationTimestamp="2026-03-18 00:21:52 +0000 UTC" firstStartedPulling="2026-03-18 00:21:54.778631072 +0000 UTC m=+832.676005672" lastFinishedPulling="2026-03-18 00:22:28.76653926 +0000 UTC m=+866.663913880" observedRunningTime="2026-03-18 00:22:30.205241991 +0000 UTC m=+868.102616611" watchObservedRunningTime="2026-03-18 00:22:30.209159192 +0000 UTC m=+868.106533802" Mar 18 00:22:30 crc kubenswrapper[4983]: I0318 00:22:30.370714 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 18 00:22:30 crc kubenswrapper[4983]: I0318 00:22:30.397486 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 18 00:22:32 crc kubenswrapper[4983]: I0318 00:22:32.182790 4983 generic.go:334] "Generic (PLEG): container finished" podID="682dc305-300d-4702-a640-b422a2185661" containerID="7cc6f1458252ccd85a9b29d27114bda72dc5482bba95dac2358c7859f23b00c5" exitCode=0 Mar 18 00:22:32 crc kubenswrapper[4983]: I0318 00:22:32.184113 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"682dc305-300d-4702-a640-b422a2185661","Type":"ContainerDied","Data":"7cc6f1458252ccd85a9b29d27114bda72dc5482bba95dac2358c7859f23b00c5"} Mar 18 00:22:35 crc kubenswrapper[4983]: I0318 00:22:35.218653 4983 generic.go:334] "Generic (PLEG): container finished" podID="682dc305-300d-4702-a640-b422a2185661" containerID="d034512699d42d5312526a8db15daf5e300a825ddb215ae6b7d3f9a30fc44d26" exitCode=0 Mar 18 00:22:35 crc kubenswrapper[4983]: I0318 00:22:35.218778 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"682dc305-300d-4702-a640-b422a2185661","Type":"ContainerDied","Data":"d034512699d42d5312526a8db15daf5e300a825ddb215ae6b7d3f9a30fc44d26"} Mar 18 00:22:35 crc kubenswrapper[4983]: I0318 00:22:35.222563 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9" event={"ID":"9e82f3b2-4502-4ea4-9fc7-55510b8e75a4","Type":"ContainerStarted","Data":"a440a364045f275d7cbb2847f026d1640a8f8989cb635ea7dbedb3d64dcb70d0"} Mar 18 00:22:35 crc kubenswrapper[4983]: I0318 00:22:35.296019 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-fjft9" podStartSLOduration=14.016200389 podStartE2EDuration="19.295993049s" podCreationTimestamp="2026-03-18 00:22:16 +0000 UTC" firstStartedPulling="2026-03-18 00:22:29.053346466 +0000 UTC m=+866.950721076" lastFinishedPulling="2026-03-18 00:22:34.333139116 +0000 UTC m=+872.230513736" observedRunningTime="2026-03-18 00:22:35.290896575 +0000 UTC m=+873.188271185" watchObservedRunningTime="2026-03-18 00:22:35.295993049 +0000 UTC m=+873.193367659" Mar 18 00:22:36 crc kubenswrapper[4983]: I0318 00:22:36.232175 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"682dc305-300d-4702-a640-b422a2185661","Type":"ContainerStarted","Data":"7080b18bcafba268ee0a2b35b526dff34797b7c8738173d33c627fb5da40d0d5"} Mar 18 00:22:36 crc kubenswrapper[4983]: I0318 00:22:36.271493 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=8.960436469 podStartE2EDuration="25.271470831s" podCreationTimestamp="2026-03-18 00:22:11 +0000 UTC" firstStartedPulling="2026-03-18 00:22:12.809133199 +0000 UTC m=+850.706507799" lastFinishedPulling="2026-03-18 00:22:29.120167551 +0000 UTC m=+867.017542161" observedRunningTime="2026-03-18 00:22:36.270547945 +0000 UTC m=+874.167922575" watchObservedRunningTime="2026-03-18 00:22:36.271470831 +0000 UTC m=+874.168845441" Mar 18 00:22:37 crc kubenswrapper[4983]: I0318 00:22:37.238409 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:37 crc kubenswrapper[4983]: I0318 00:22:37.899222 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-f7fhd"] Mar 18 00:22:37 crc kubenswrapper[4983]: I0318 00:22:37.899987 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-f7fhd" Mar 18 00:22:37 crc kubenswrapper[4983]: I0318 00:22:37.905006 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 18 00:22:37 crc kubenswrapper[4983]: I0318 00:22:37.905053 4983 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-c5ld7" Mar 18 00:22:37 crc kubenswrapper[4983]: I0318 00:22:37.906534 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 18 00:22:37 crc kubenswrapper[4983]: I0318 00:22:37.912984 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-f7fhd"] Mar 18 00:22:38 crc kubenswrapper[4983]: I0318 00:22:38.048614 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2bpm\" (UniqueName: \"kubernetes.io/projected/3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2-kube-api-access-s2bpm\") pod \"cert-manager-webhook-6888856db4-f7fhd\" (UID: \"3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2\") " pod="cert-manager/cert-manager-webhook-6888856db4-f7fhd" Mar 18 00:22:38 crc kubenswrapper[4983]: I0318 00:22:38.048698 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-f7fhd\" (UID: \"3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2\") " pod="cert-manager/cert-manager-webhook-6888856db4-f7fhd" Mar 18 00:22:38 crc kubenswrapper[4983]: I0318 00:22:38.150009 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-f7fhd\" (UID: \"3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2\") " pod="cert-manager/cert-manager-webhook-6888856db4-f7fhd" Mar 18 00:22:38 crc kubenswrapper[4983]: I0318 00:22:38.150464 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2bpm\" (UniqueName: \"kubernetes.io/projected/3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2-kube-api-access-s2bpm\") pod \"cert-manager-webhook-6888856db4-f7fhd\" (UID: \"3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2\") " pod="cert-manager/cert-manager-webhook-6888856db4-f7fhd" Mar 18 00:22:38 crc kubenswrapper[4983]: I0318 00:22:38.175788 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2bpm\" (UniqueName: \"kubernetes.io/projected/3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2-kube-api-access-s2bpm\") pod \"cert-manager-webhook-6888856db4-f7fhd\" (UID: \"3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2\") " pod="cert-manager/cert-manager-webhook-6888856db4-f7fhd" Mar 18 00:22:38 crc kubenswrapper[4983]: I0318 00:22:38.179215 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-f7fhd\" (UID: \"3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2\") " pod="cert-manager/cert-manager-webhook-6888856db4-f7fhd" Mar 18 00:22:38 crc kubenswrapper[4983]: I0318 00:22:38.214237 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-f7fhd" Mar 18 00:22:38 crc kubenswrapper[4983]: I0318 00:22:38.580773 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-f7fhd"] Mar 18 00:22:38 crc kubenswrapper[4983]: W0318 00:22:38.591099 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f721b9c_8bd6_4e8c_b549_fa5ed88a92b2.slice/crio-93d97fc3a009625b232d6b255cc90c06fdac510cf9f9ea4aa3d584a8dcee7506 WatchSource:0}: Error finding container 93d97fc3a009625b232d6b255cc90c06fdac510cf9f9ea4aa3d584a8dcee7506: Status 404 returned error can't find the container with id 93d97fc3a009625b232d6b255cc90c06fdac510cf9f9ea4aa3d584a8dcee7506 Mar 18 00:22:39 crc kubenswrapper[4983]: I0318 00:22:39.251570 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-f7fhd" event={"ID":"3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2","Type":"ContainerStarted","Data":"93d97fc3a009625b232d6b255cc90c06fdac510cf9f9ea4aa3d584a8dcee7506"} Mar 18 00:22:41 crc kubenswrapper[4983]: I0318 00:22:41.415491 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-2xxpj"] Mar 18 00:22:41 crc kubenswrapper[4983]: I0318 00:22:41.417437 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-2xxpj" Mar 18 00:22:41 crc kubenswrapper[4983]: I0318 00:22:41.422638 4983 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-p846h" Mar 18 00:22:41 crc kubenswrapper[4983]: I0318 00:22:41.431648 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-2xxpj"] Mar 18 00:22:41 crc kubenswrapper[4983]: I0318 00:22:41.490290 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/726bff1a-cb5b-495b-aa52-61500dbc5616-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-2xxpj\" (UID: \"726bff1a-cb5b-495b-aa52-61500dbc5616\") " pod="cert-manager/cert-manager-cainjector-5545bd876-2xxpj" Mar 18 00:22:41 crc kubenswrapper[4983]: I0318 00:22:41.490366 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdgsx\" (UniqueName: \"kubernetes.io/projected/726bff1a-cb5b-495b-aa52-61500dbc5616-kube-api-access-pdgsx\") pod \"cert-manager-cainjector-5545bd876-2xxpj\" (UID: \"726bff1a-cb5b-495b-aa52-61500dbc5616\") " pod="cert-manager/cert-manager-cainjector-5545bd876-2xxpj" Mar 18 00:22:41 crc kubenswrapper[4983]: I0318 00:22:41.591882 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/726bff1a-cb5b-495b-aa52-61500dbc5616-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-2xxpj\" (UID: \"726bff1a-cb5b-495b-aa52-61500dbc5616\") " pod="cert-manager/cert-manager-cainjector-5545bd876-2xxpj" Mar 18 00:22:41 crc kubenswrapper[4983]: I0318 00:22:41.592071 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdgsx\" (UniqueName: \"kubernetes.io/projected/726bff1a-cb5b-495b-aa52-61500dbc5616-kube-api-access-pdgsx\") pod \"cert-manager-cainjector-5545bd876-2xxpj\" (UID: \"726bff1a-cb5b-495b-aa52-61500dbc5616\") " pod="cert-manager/cert-manager-cainjector-5545bd876-2xxpj" Mar 18 00:22:41 crc kubenswrapper[4983]: I0318 00:22:41.617296 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/726bff1a-cb5b-495b-aa52-61500dbc5616-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-2xxpj\" (UID: \"726bff1a-cb5b-495b-aa52-61500dbc5616\") " pod="cert-manager/cert-manager-cainjector-5545bd876-2xxpj" Mar 18 00:22:41 crc kubenswrapper[4983]: I0318 00:22:41.628603 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdgsx\" (UniqueName: \"kubernetes.io/projected/726bff1a-cb5b-495b-aa52-61500dbc5616-kube-api-access-pdgsx\") pod \"cert-manager-cainjector-5545bd876-2xxpj\" (UID: \"726bff1a-cb5b-495b-aa52-61500dbc5616\") " pod="cert-manager/cert-manager-cainjector-5545bd876-2xxpj" Mar 18 00:22:41 crc kubenswrapper[4983]: I0318 00:22:41.734436 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-2xxpj" Mar 18 00:22:42 crc kubenswrapper[4983]: I0318 00:22:42.212503 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-2xxpj"] Mar 18 00:22:42 crc kubenswrapper[4983]: W0318 00:22:42.217500 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod726bff1a_cb5b_495b_aa52_61500dbc5616.slice/crio-13035a6f7c120242efa8300dfaaf9eb83de9111001ef8b9c8d3415bd11e8ba7f WatchSource:0}: Error finding container 13035a6f7c120242efa8300dfaaf9eb83de9111001ef8b9c8d3415bd11e8ba7f: Status 404 returned error can't find the container with id 13035a6f7c120242efa8300dfaaf9eb83de9111001ef8b9c8d3415bd11e8ba7f Mar 18 00:22:42 crc kubenswrapper[4983]: I0318 00:22:42.288777 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-2xxpj" event={"ID":"726bff1a-cb5b-495b-aa52-61500dbc5616","Type":"ContainerStarted","Data":"13035a6f7c120242efa8300dfaaf9eb83de9111001ef8b9c8d3415bd11e8ba7f"} Mar 18 00:22:43 crc kubenswrapper[4983]: I0318 00:22:43.123355 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-749b9786b8-rccft" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.397539 4983 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="682dc305-300d-4702-a640-b422a2185661" containerName="elasticsearch" probeResult="failure" output=< Mar 18 00:22:47 crc kubenswrapper[4983]: {"timestamp": "2026-03-18T00:22:47+00:00", "message": "readiness probe failed", "curl_rc": "7"} Mar 18 00:22:47 crc kubenswrapper[4983]: > Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.402415 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.403381 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.407467 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-q2gzr" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.409655 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-sys-config" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.409697 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-global-ca" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.409700 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-ca" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.428908 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.578417 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.578483 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/11bca75d-c62d-45d8-9076-18f5a1d79715-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.578508 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.578528 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.578611 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/11bca75d-c62d-45d8-9076-18f5a1d79715-builder-dockercfg-q2gzr-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.578650 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.578823 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.578902 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/11bca75d-c62d-45d8-9076-18f5a1d79715-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.578942 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/11bca75d-c62d-45d8-9076-18f5a1d79715-builder-dockercfg-q2gzr-push\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.578965 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwbdt\" (UniqueName: \"kubernetes.io/projected/11bca75d-c62d-45d8-9076-18f5a1d79715-kube-api-access-jwbdt\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.578986 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.579002 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.680728 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.680807 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/11bca75d-c62d-45d8-9076-18f5a1d79715-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.680842 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.680859 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.680879 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/11bca75d-c62d-45d8-9076-18f5a1d79715-builder-dockercfg-q2gzr-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.680894 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.680945 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.680939 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/11bca75d-c62d-45d8-9076-18f5a1d79715-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.680971 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/11bca75d-c62d-45d8-9076-18f5a1d79715-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.681019 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/11bca75d-c62d-45d8-9076-18f5a1d79715-builder-dockercfg-q2gzr-push\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.681041 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/11bca75d-c62d-45d8-9076-18f5a1d79715-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.681039 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwbdt\" (UniqueName: \"kubernetes.io/projected/11bca75d-c62d-45d8-9076-18f5a1d79715-kube-api-access-jwbdt\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.681084 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.681112 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.681989 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.682218 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.682233 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.682253 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.682713 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.682781 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.683011 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.686998 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/11bca75d-c62d-45d8-9076-18f5a1d79715-builder-dockercfg-q2gzr-push\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.687395 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/11bca75d-c62d-45d8-9076-18f5a1d79715-builder-dockercfg-q2gzr-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:47 crc kubenswrapper[4983]: I0318 00:22:47.719574 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwbdt\" (UniqueName: \"kubernetes.io/projected/11bca75d-c62d-45d8-9076-18f5a1d79715-kube-api-access-jwbdt\") pod \"service-telemetry-operator-1-build\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:48 crc kubenswrapper[4983]: I0318 00:22:48.018232 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.216801 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.328457 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"11bca75d-c62d-45d8-9076-18f5a1d79715","Type":"ContainerStarted","Data":"0da1ef9f6a210406932fce7fdd8c04e2937f01f9f032fa433f34fb00d179d567"} Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.329666 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-2xxpj" event={"ID":"726bff1a-cb5b-495b-aa52-61500dbc5616","Type":"ContainerStarted","Data":"c603d8e32afb792a9b3a8c02897b0dbf36e462b3d04f501ede1ed61a4cb54fd1"} Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.330873 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-f7fhd" event={"ID":"3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2","Type":"ContainerStarted","Data":"b582a13b630edfa8312fa4c9fcb8a277a368c7182e86281dc1159eec3a1e6ef4"} Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.331076 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-f7fhd" Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.439879 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-2xxpj" podStartSLOduration=1.678021494 podStartE2EDuration="8.439858486s" podCreationTimestamp="2026-03-18 00:22:41 +0000 UTC" firstStartedPulling="2026-03-18 00:22:42.220842565 +0000 UTC m=+880.118217175" lastFinishedPulling="2026-03-18 00:22:48.982679557 +0000 UTC m=+886.880054167" observedRunningTime="2026-03-18 00:22:49.360508595 +0000 UTC m=+887.257883205" watchObservedRunningTime="2026-03-18 00:22:49.439858486 +0000 UTC m=+887.337233096" Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.443880 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-lfdvf"] Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.444686 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-lfdvf" Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.448659 4983 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-st6v2" Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.471789 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-f7fhd" podStartSLOduration=2.107695995 podStartE2EDuration="12.471770881s" podCreationTimestamp="2026-03-18 00:22:37 +0000 UTC" firstStartedPulling="2026-03-18 00:22:38.601362742 +0000 UTC m=+876.498737342" lastFinishedPulling="2026-03-18 00:22:48.965437618 +0000 UTC m=+886.862812228" observedRunningTime="2026-03-18 00:22:49.470113774 +0000 UTC m=+887.367488384" watchObservedRunningTime="2026-03-18 00:22:49.471770881 +0000 UTC m=+887.369145491" Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.481102 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-lfdvf"] Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.515215 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnfcm\" (UniqueName: \"kubernetes.io/projected/7d75c786-fad6-40d0-85e6-799f383cb4ba-kube-api-access-vnfcm\") pod \"cert-manager-545d4d4674-lfdvf\" (UID: \"7d75c786-fad6-40d0-85e6-799f383cb4ba\") " pod="cert-manager/cert-manager-545d4d4674-lfdvf" Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.515336 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7d75c786-fad6-40d0-85e6-799f383cb4ba-bound-sa-token\") pod \"cert-manager-545d4d4674-lfdvf\" (UID: \"7d75c786-fad6-40d0-85e6-799f383cb4ba\") " pod="cert-manager/cert-manager-545d4d4674-lfdvf" Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.616135 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7d75c786-fad6-40d0-85e6-799f383cb4ba-bound-sa-token\") pod \"cert-manager-545d4d4674-lfdvf\" (UID: \"7d75c786-fad6-40d0-85e6-799f383cb4ba\") " pod="cert-manager/cert-manager-545d4d4674-lfdvf" Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.616223 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnfcm\" (UniqueName: \"kubernetes.io/projected/7d75c786-fad6-40d0-85e6-799f383cb4ba-kube-api-access-vnfcm\") pod \"cert-manager-545d4d4674-lfdvf\" (UID: \"7d75c786-fad6-40d0-85e6-799f383cb4ba\") " pod="cert-manager/cert-manager-545d4d4674-lfdvf" Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.640728 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7d75c786-fad6-40d0-85e6-799f383cb4ba-bound-sa-token\") pod \"cert-manager-545d4d4674-lfdvf\" (UID: \"7d75c786-fad6-40d0-85e6-799f383cb4ba\") " pod="cert-manager/cert-manager-545d4d4674-lfdvf" Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.656760 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnfcm\" (UniqueName: \"kubernetes.io/projected/7d75c786-fad6-40d0-85e6-799f383cb4ba-kube-api-access-vnfcm\") pod \"cert-manager-545d4d4674-lfdvf\" (UID: \"7d75c786-fad6-40d0-85e6-799f383cb4ba\") " pod="cert-manager/cert-manager-545d4d4674-lfdvf" Mar 18 00:22:49 crc kubenswrapper[4983]: I0318 00:22:49.759355 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-lfdvf" Mar 18 00:22:50 crc kubenswrapper[4983]: I0318 00:22:50.345638 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-lfdvf"] Mar 18 00:22:50 crc kubenswrapper[4983]: W0318 00:22:50.362730 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d75c786_fad6_40d0_85e6_799f383cb4ba.slice/crio-4cc14fcd1aa8fdf9e8fa2275105e8eb24f499b67ad77a2abb5adfbb466ffc1af WatchSource:0}: Error finding container 4cc14fcd1aa8fdf9e8fa2275105e8eb24f499b67ad77a2abb5adfbb466ffc1af: Status 404 returned error can't find the container with id 4cc14fcd1aa8fdf9e8fa2275105e8eb24f499b67ad77a2abb5adfbb466ffc1af Mar 18 00:22:51 crc kubenswrapper[4983]: I0318 00:22:51.363307 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-lfdvf" event={"ID":"7d75c786-fad6-40d0-85e6-799f383cb4ba","Type":"ContainerStarted","Data":"e07213bc553aaac94d5fc27f5b1445e54706dc817b22e1c43041f1203d06e2bb"} Mar 18 00:22:51 crc kubenswrapper[4983]: I0318 00:22:51.363625 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-lfdvf" event={"ID":"7d75c786-fad6-40d0-85e6-799f383cb4ba","Type":"ContainerStarted","Data":"4cc14fcd1aa8fdf9e8fa2275105e8eb24f499b67ad77a2abb5adfbb466ffc1af"} Mar 18 00:22:51 crc kubenswrapper[4983]: I0318 00:22:51.390997 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-lfdvf" podStartSLOduration=2.390970522 podStartE2EDuration="2.390970522s" podCreationTimestamp="2026-03-18 00:22:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:22:51.3905388 +0000 UTC m=+889.287913410" watchObservedRunningTime="2026-03-18 00:22:51.390970522 +0000 UTC m=+889.288345132" Mar 18 00:22:52 crc kubenswrapper[4983]: I0318 00:22:52.704572 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Mar 18 00:22:57 crc kubenswrapper[4983]: I0318 00:22:57.377994 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.218361 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-f7fhd" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.418191 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"11bca75d-c62d-45d8-9076-18f5a1d79715","Type":"ContainerStarted","Data":"64f0f291faf4d7ed64038ccd91ecb6b5afee49fbcb00294b452e1aee9a997b79"} Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.418413 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-1-build" podUID="11bca75d-c62d-45d8-9076-18f5a1d79715" containerName="manage-dockerfile" containerID="cri-o://64f0f291faf4d7ed64038ccd91ecb6b5afee49fbcb00294b452e1aee9a997b79" gracePeriod=30 Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.834780 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_11bca75d-c62d-45d8-9076-18f5a1d79715/manage-dockerfile/0.log" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.835143 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.960255 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/11bca75d-c62d-45d8-9076-18f5a1d79715-buildcachedir\") pod \"11bca75d-c62d-45d8-9076-18f5a1d79715\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.960308 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-ca-bundles\") pod \"11bca75d-c62d-45d8-9076-18f5a1d79715\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.960337 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/11bca75d-c62d-45d8-9076-18f5a1d79715-node-pullsecrets\") pod \"11bca75d-c62d-45d8-9076-18f5a1d79715\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.960359 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-buildworkdir\") pod \"11bca75d-c62d-45d8-9076-18f5a1d79715\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.960384 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwbdt\" (UniqueName: \"kubernetes.io/projected/11bca75d-c62d-45d8-9076-18f5a1d79715-kube-api-access-jwbdt\") pod \"11bca75d-c62d-45d8-9076-18f5a1d79715\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.960372 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11bca75d-c62d-45d8-9076-18f5a1d79715-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "11bca75d-c62d-45d8-9076-18f5a1d79715" (UID: "11bca75d-c62d-45d8-9076-18f5a1d79715"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.960434 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-container-storage-root\") pod \"11bca75d-c62d-45d8-9076-18f5a1d79715\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.960800 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/11bca75d-c62d-45d8-9076-18f5a1d79715-builder-dockercfg-q2gzr-push\") pod \"11bca75d-c62d-45d8-9076-18f5a1d79715\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.960480 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11bca75d-c62d-45d8-9076-18f5a1d79715-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "11bca75d-c62d-45d8-9076-18f5a1d79715" (UID: "11bca75d-c62d-45d8-9076-18f5a1d79715"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.960738 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "11bca75d-c62d-45d8-9076-18f5a1d79715" (UID: "11bca75d-c62d-45d8-9076-18f5a1d79715"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.960886 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-proxy-ca-bundles\") pod \"11bca75d-c62d-45d8-9076-18f5a1d79715\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.960905 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-system-configs\") pod \"11bca75d-c62d-45d8-9076-18f5a1d79715\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.960972 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "11bca75d-c62d-45d8-9076-18f5a1d79715" (UID: "11bca75d-c62d-45d8-9076-18f5a1d79715"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961350 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "11bca75d-c62d-45d8-9076-18f5a1d79715" (UID: "11bca75d-c62d-45d8-9076-18f5a1d79715"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961363 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "11bca75d-c62d-45d8-9076-18f5a1d79715" (UID: "11bca75d-c62d-45d8-9076-18f5a1d79715"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961389 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/11bca75d-c62d-45d8-9076-18f5a1d79715-builder-dockercfg-q2gzr-pull\") pod \"11bca75d-c62d-45d8-9076-18f5a1d79715\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961414 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-container-storage-run\") pod \"11bca75d-c62d-45d8-9076-18f5a1d79715\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961441 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-build-blob-cache\") pod \"11bca75d-c62d-45d8-9076-18f5a1d79715\" (UID: \"11bca75d-c62d-45d8-9076-18f5a1d79715\") " Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961600 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "11bca75d-c62d-45d8-9076-18f5a1d79715" (UID: "11bca75d-c62d-45d8-9076-18f5a1d79715"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961688 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "11bca75d-c62d-45d8-9076-18f5a1d79715" (UID: "11bca75d-c62d-45d8-9076-18f5a1d79715"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961757 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "11bca75d-c62d-45d8-9076-18f5a1d79715" (UID: "11bca75d-c62d-45d8-9076-18f5a1d79715"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961830 4983 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/11bca75d-c62d-45d8-9076-18f5a1d79715-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961846 4983 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961856 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961869 4983 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961877 4983 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961885 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.961895 4983 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/11bca75d-c62d-45d8-9076-18f5a1d79715-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.966059 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11bca75d-c62d-45d8-9076-18f5a1d79715-builder-dockercfg-q2gzr-push" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-push") pod "11bca75d-c62d-45d8-9076-18f5a1d79715" (UID: "11bca75d-c62d-45d8-9076-18f5a1d79715"). InnerVolumeSpecName "builder-dockercfg-q2gzr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.967071 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11bca75d-c62d-45d8-9076-18f5a1d79715-kube-api-access-jwbdt" (OuterVolumeSpecName: "kube-api-access-jwbdt") pod "11bca75d-c62d-45d8-9076-18f5a1d79715" (UID: "11bca75d-c62d-45d8-9076-18f5a1d79715"). InnerVolumeSpecName "kube-api-access-jwbdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:22:58 crc kubenswrapper[4983]: I0318 00:22:58.969096 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11bca75d-c62d-45d8-9076-18f5a1d79715-builder-dockercfg-q2gzr-pull" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-pull") pod "11bca75d-c62d-45d8-9076-18f5a1d79715" (UID: "11bca75d-c62d-45d8-9076-18f5a1d79715"). InnerVolumeSpecName "builder-dockercfg-q2gzr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.059069 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Mar 18 00:22:59 crc kubenswrapper[4983]: E0318 00:22:59.059318 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11bca75d-c62d-45d8-9076-18f5a1d79715" containerName="manage-dockerfile" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.059330 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="11bca75d-c62d-45d8-9076-18f5a1d79715" containerName="manage-dockerfile" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.059431 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="11bca75d-c62d-45d8-9076-18f5a1d79715" containerName="manage-dockerfile" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.060247 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.062456 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-ca" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.062634 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-global-ca" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.062842 4983 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/11bca75d-c62d-45d8-9076-18f5a1d79715-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.062874 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwbdt\" (UniqueName: \"kubernetes.io/projected/11bca75d-c62d-45d8-9076-18f5a1d79715-kube-api-access-jwbdt\") on node \"crc\" DevicePath \"\"" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.062888 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/11bca75d-c62d-45d8-9076-18f5a1d79715-builder-dockercfg-q2gzr-push\") on node \"crc\" DevicePath \"\"" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.062904 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/11bca75d-c62d-45d8-9076-18f5a1d79715-builder-dockercfg-q2gzr-pull\") on node \"crc\" DevicePath \"\"" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.062916 4983 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/11bca75d-c62d-45d8-9076-18f5a1d79715-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.062944 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-sys-config" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.088329 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.164610 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/493c24a2-c93e-481a-b2c4-61254df38d13-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.164901 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.164951 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/493c24a2-c93e-481a-b2c4-61254df38d13-builder-dockercfg-q2gzr-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.164974 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/493c24a2-c93e-481a-b2c4-61254df38d13-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.165007 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.165257 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfn4x\" (UniqueName: \"kubernetes.io/projected/493c24a2-c93e-481a-b2c4-61254df38d13-kube-api-access-sfn4x\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.165283 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.165481 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.165591 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.165620 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/493c24a2-c93e-481a-b2c4-61254df38d13-builder-dockercfg-q2gzr-push\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.165673 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.165705 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.267019 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.267106 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.267140 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/493c24a2-c93e-481a-b2c4-61254df38d13-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.267180 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.267220 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/493c24a2-c93e-481a-b2c4-61254df38d13-builder-dockercfg-q2gzr-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.267246 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/493c24a2-c93e-481a-b2c4-61254df38d13-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.267277 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.267298 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfn4x\" (UniqueName: \"kubernetes.io/projected/493c24a2-c93e-481a-b2c4-61254df38d13-kube-api-access-sfn4x\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.267321 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/493c24a2-c93e-481a-b2c4-61254df38d13-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.267379 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/493c24a2-c93e-481a-b2c4-61254df38d13-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.267849 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.267901 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.268167 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.268235 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.268326 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.268357 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.268455 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.269025 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.269133 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.269472 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.268487 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/493c24a2-c93e-481a-b2c4-61254df38d13-builder-dockercfg-q2gzr-push\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.276979 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/493c24a2-c93e-481a-b2c4-61254df38d13-builder-dockercfg-q2gzr-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.287317 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/493c24a2-c93e-481a-b2c4-61254df38d13-builder-dockercfg-q2gzr-push\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.290535 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfn4x\" (UniqueName: \"kubernetes.io/projected/493c24a2-c93e-481a-b2c4-61254df38d13-kube-api-access-sfn4x\") pod \"service-telemetry-operator-2-build\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.374496 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.454496 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_11bca75d-c62d-45d8-9076-18f5a1d79715/manage-dockerfile/0.log" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.454813 4983 generic.go:334] "Generic (PLEG): container finished" podID="11bca75d-c62d-45d8-9076-18f5a1d79715" containerID="64f0f291faf4d7ed64038ccd91ecb6b5afee49fbcb00294b452e1aee9a997b79" exitCode=1 Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.454853 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"11bca75d-c62d-45d8-9076-18f5a1d79715","Type":"ContainerDied","Data":"64f0f291faf4d7ed64038ccd91ecb6b5afee49fbcb00294b452e1aee9a997b79"} Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.454889 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"11bca75d-c62d-45d8-9076-18f5a1d79715","Type":"ContainerDied","Data":"0da1ef9f6a210406932fce7fdd8c04e2937f01f9f032fa433f34fb00d179d567"} Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.454907 4983 scope.go:117] "RemoveContainer" containerID="64f0f291faf4d7ed64038ccd91ecb6b5afee49fbcb00294b452e1aee9a997b79" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.454921 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.480899 4983 scope.go:117] "RemoveContainer" containerID="64f0f291faf4d7ed64038ccd91ecb6b5afee49fbcb00294b452e1aee9a997b79" Mar 18 00:22:59 crc kubenswrapper[4983]: E0318 00:22:59.481334 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64f0f291faf4d7ed64038ccd91ecb6b5afee49fbcb00294b452e1aee9a997b79\": container with ID starting with 64f0f291faf4d7ed64038ccd91ecb6b5afee49fbcb00294b452e1aee9a997b79 not found: ID does not exist" containerID="64f0f291faf4d7ed64038ccd91ecb6b5afee49fbcb00294b452e1aee9a997b79" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.481400 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64f0f291faf4d7ed64038ccd91ecb6b5afee49fbcb00294b452e1aee9a997b79"} err="failed to get container status \"64f0f291faf4d7ed64038ccd91ecb6b5afee49fbcb00294b452e1aee9a997b79\": rpc error: code = NotFound desc = could not find container \"64f0f291faf4d7ed64038ccd91ecb6b5afee49fbcb00294b452e1aee9a997b79\": container with ID starting with 64f0f291faf4d7ed64038ccd91ecb6b5afee49fbcb00294b452e1aee9a997b79 not found: ID does not exist" Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.491170 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.496821 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 18 00:22:59 crc kubenswrapper[4983]: I0318 00:22:59.691166 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Mar 18 00:23:00 crc kubenswrapper[4983]: I0318 00:23:00.463308 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"493c24a2-c93e-481a-b2c4-61254df38d13","Type":"ContainerStarted","Data":"64355c1b7cc2d275bc6d7dcf29065dbb15616fc89a72be9d8c8611022dba70ba"} Mar 18 00:23:00 crc kubenswrapper[4983]: I0318 00:23:00.463360 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"493c24a2-c93e-481a-b2c4-61254df38d13","Type":"ContainerStarted","Data":"bef5ca5e98567d173e7048df62149130324df97f1e1a252e909ed0bb77f0d4b8"} Mar 18 00:23:00 crc kubenswrapper[4983]: I0318 00:23:00.868715 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11bca75d-c62d-45d8-9076-18f5a1d79715" path="/var/lib/kubelet/pods/11bca75d-c62d-45d8-9076-18f5a1d79715/volumes" Mar 18 00:23:06 crc kubenswrapper[4983]: I0318 00:23:06.525698 4983 generic.go:334] "Generic (PLEG): container finished" podID="493c24a2-c93e-481a-b2c4-61254df38d13" containerID="64355c1b7cc2d275bc6d7dcf29065dbb15616fc89a72be9d8c8611022dba70ba" exitCode=0 Mar 18 00:23:06 crc kubenswrapper[4983]: I0318 00:23:06.525886 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"493c24a2-c93e-481a-b2c4-61254df38d13","Type":"ContainerDied","Data":"64355c1b7cc2d275bc6d7dcf29065dbb15616fc89a72be9d8c8611022dba70ba"} Mar 18 00:23:07 crc kubenswrapper[4983]: I0318 00:23:07.536206 4983 generic.go:334] "Generic (PLEG): container finished" podID="493c24a2-c93e-481a-b2c4-61254df38d13" containerID="99a68dc4d8cf12af96d32badf816f249d8a0db1b07678b0dd6139157ff51dfee" exitCode=0 Mar 18 00:23:07 crc kubenswrapper[4983]: I0318 00:23:07.536275 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"493c24a2-c93e-481a-b2c4-61254df38d13","Type":"ContainerDied","Data":"99a68dc4d8cf12af96d32badf816f249d8a0db1b07678b0dd6139157ff51dfee"} Mar 18 00:23:07 crc kubenswrapper[4983]: I0318 00:23:07.584800 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_493c24a2-c93e-481a-b2c4-61254df38d13/manage-dockerfile/0.log" Mar 18 00:23:08 crc kubenswrapper[4983]: I0318 00:23:08.565531 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"493c24a2-c93e-481a-b2c4-61254df38d13","Type":"ContainerStarted","Data":"1bcd4f1537989b9603ac254a672e191249cbd5ec72167fd79a9bb6fc4b18fe69"} Mar 18 00:23:08 crc kubenswrapper[4983]: I0318 00:23:08.630713 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-2-build" podStartSLOduration=9.630681608 podStartE2EDuration="9.630681608s" podCreationTimestamp="2026-03-18 00:22:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:23:08.62864083 +0000 UTC m=+906.526015510" watchObservedRunningTime="2026-03-18 00:23:08.630681608 +0000 UTC m=+906.528056268" Mar 18 00:23:09 crc kubenswrapper[4983]: I0318 00:23:09.693877 4983 scope.go:117] "RemoveContainer" containerID="1214a03f220209d0bfd44f9dc961ae1622eb933596e325b50d76eb0c01361fb8" Mar 18 00:23:43 crc kubenswrapper[4983]: I0318 00:23:43.259554 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:23:43 crc kubenswrapper[4983]: I0318 00:23:43.260362 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:24:00 crc kubenswrapper[4983]: I0318 00:24:00.139620 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563224-b4g22"] Mar 18 00:24:00 crc kubenswrapper[4983]: I0318 00:24:00.141325 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563224-b4g22" Mar 18 00:24:00 crc kubenswrapper[4983]: I0318 00:24:00.144164 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:24:00 crc kubenswrapper[4983]: I0318 00:24:00.144283 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:24:00 crc kubenswrapper[4983]: I0318 00:24:00.146672 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:24:00 crc kubenswrapper[4983]: I0318 00:24:00.150854 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563224-b4g22"] Mar 18 00:24:00 crc kubenswrapper[4983]: I0318 00:24:00.273287 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq2m6\" (UniqueName: \"kubernetes.io/projected/15371e16-94d2-4fa9-a891-6168661fdaf7-kube-api-access-fq2m6\") pod \"auto-csr-approver-29563224-b4g22\" (UID: \"15371e16-94d2-4fa9-a891-6168661fdaf7\") " pod="openshift-infra/auto-csr-approver-29563224-b4g22" Mar 18 00:24:00 crc kubenswrapper[4983]: I0318 00:24:00.374394 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq2m6\" (UniqueName: \"kubernetes.io/projected/15371e16-94d2-4fa9-a891-6168661fdaf7-kube-api-access-fq2m6\") pod \"auto-csr-approver-29563224-b4g22\" (UID: \"15371e16-94d2-4fa9-a891-6168661fdaf7\") " pod="openshift-infra/auto-csr-approver-29563224-b4g22" Mar 18 00:24:00 crc kubenswrapper[4983]: I0318 00:24:00.393382 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq2m6\" (UniqueName: \"kubernetes.io/projected/15371e16-94d2-4fa9-a891-6168661fdaf7-kube-api-access-fq2m6\") pod \"auto-csr-approver-29563224-b4g22\" (UID: \"15371e16-94d2-4fa9-a891-6168661fdaf7\") " pod="openshift-infra/auto-csr-approver-29563224-b4g22" Mar 18 00:24:00 crc kubenswrapper[4983]: I0318 00:24:00.458836 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563224-b4g22" Mar 18 00:24:00 crc kubenswrapper[4983]: I0318 00:24:00.918676 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563224-b4g22"] Mar 18 00:24:00 crc kubenswrapper[4983]: W0318 00:24:00.941858 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15371e16_94d2_4fa9_a891_6168661fdaf7.slice/crio-9a8d518a48aceafc25e4aba7c632a5b64a57d2c783658e9e84b2817cdf46dfac WatchSource:0}: Error finding container 9a8d518a48aceafc25e4aba7c632a5b64a57d2c783658e9e84b2817cdf46dfac: Status 404 returned error can't find the container with id 9a8d518a48aceafc25e4aba7c632a5b64a57d2c783658e9e84b2817cdf46dfac Mar 18 00:24:00 crc kubenswrapper[4983]: I0318 00:24:00.944546 4983 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 00:24:01 crc kubenswrapper[4983]: I0318 00:24:01.934058 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563224-b4g22" event={"ID":"15371e16-94d2-4fa9-a891-6168661fdaf7","Type":"ContainerStarted","Data":"9a8d518a48aceafc25e4aba7c632a5b64a57d2c783658e9e84b2817cdf46dfac"} Mar 18 00:24:02 crc kubenswrapper[4983]: I0318 00:24:02.944371 4983 generic.go:334] "Generic (PLEG): container finished" podID="15371e16-94d2-4fa9-a891-6168661fdaf7" containerID="e41b566259dca5a930a81be7a73d899985af9103788d9beb47f5b570ed6dfd52" exitCode=0 Mar 18 00:24:02 crc kubenswrapper[4983]: I0318 00:24:02.945063 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563224-b4g22" event={"ID":"15371e16-94d2-4fa9-a891-6168661fdaf7","Type":"ContainerDied","Data":"e41b566259dca5a930a81be7a73d899985af9103788d9beb47f5b570ed6dfd52"} Mar 18 00:24:04 crc kubenswrapper[4983]: I0318 00:24:04.186654 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563224-b4g22" Mar 18 00:24:04 crc kubenswrapper[4983]: I0318 00:24:04.345194 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fq2m6\" (UniqueName: \"kubernetes.io/projected/15371e16-94d2-4fa9-a891-6168661fdaf7-kube-api-access-fq2m6\") pod \"15371e16-94d2-4fa9-a891-6168661fdaf7\" (UID: \"15371e16-94d2-4fa9-a891-6168661fdaf7\") " Mar 18 00:24:04 crc kubenswrapper[4983]: I0318 00:24:04.350719 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15371e16-94d2-4fa9-a891-6168661fdaf7-kube-api-access-fq2m6" (OuterVolumeSpecName: "kube-api-access-fq2m6") pod "15371e16-94d2-4fa9-a891-6168661fdaf7" (UID: "15371e16-94d2-4fa9-a891-6168661fdaf7"). InnerVolumeSpecName "kube-api-access-fq2m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:24:04 crc kubenswrapper[4983]: I0318 00:24:04.447707 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fq2m6\" (UniqueName: \"kubernetes.io/projected/15371e16-94d2-4fa9-a891-6168661fdaf7-kube-api-access-fq2m6\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:04 crc kubenswrapper[4983]: I0318 00:24:04.966157 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563224-b4g22" event={"ID":"15371e16-94d2-4fa9-a891-6168661fdaf7","Type":"ContainerDied","Data":"9a8d518a48aceafc25e4aba7c632a5b64a57d2c783658e9e84b2817cdf46dfac"} Mar 18 00:24:04 crc kubenswrapper[4983]: I0318 00:24:04.966210 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a8d518a48aceafc25e4aba7c632a5b64a57d2c783658e9e84b2817cdf46dfac" Mar 18 00:24:04 crc kubenswrapper[4983]: I0318 00:24:04.966473 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563224-b4g22" Mar 18 00:24:05 crc kubenswrapper[4983]: I0318 00:24:05.253576 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563218-f2zls"] Mar 18 00:24:05 crc kubenswrapper[4983]: I0318 00:24:05.260004 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563218-f2zls"] Mar 18 00:24:06 crc kubenswrapper[4983]: I0318 00:24:06.857915 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97fb5d36-22f5-493c-a22f-e857b801905c" path="/var/lib/kubelet/pods/97fb5d36-22f5-493c-a22f-e857b801905c/volumes" Mar 18 00:24:09 crc kubenswrapper[4983]: I0318 00:24:09.795433 4983 scope.go:117] "RemoveContainer" containerID="89ddb68660edebbbebab0a216bba5801a0590e6e8d8d43fff9058d340c069d97" Mar 18 00:24:13 crc kubenswrapper[4983]: I0318 00:24:13.260101 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:24:13 crc kubenswrapper[4983]: I0318 00:24:13.260999 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.562411 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dlpmx"] Mar 18 00:24:23 crc kubenswrapper[4983]: E0318 00:24:23.563105 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15371e16-94d2-4fa9-a891-6168661fdaf7" containerName="oc" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.563123 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="15371e16-94d2-4fa9-a891-6168661fdaf7" containerName="oc" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.563270 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="15371e16-94d2-4fa9-a891-6168661fdaf7" containerName="oc" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.564434 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.579677 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dlpmx"] Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.656827 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71e0f21c-fafb-41de-9726-4a840f0a7599-utilities\") pod \"redhat-operators-dlpmx\" (UID: \"71e0f21c-fafb-41de-9726-4a840f0a7599\") " pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.656886 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71e0f21c-fafb-41de-9726-4a840f0a7599-catalog-content\") pod \"redhat-operators-dlpmx\" (UID: \"71e0f21c-fafb-41de-9726-4a840f0a7599\") " pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.656921 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wfqb\" (UniqueName: \"kubernetes.io/projected/71e0f21c-fafb-41de-9726-4a840f0a7599-kube-api-access-6wfqb\") pod \"redhat-operators-dlpmx\" (UID: \"71e0f21c-fafb-41de-9726-4a840f0a7599\") " pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.758984 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71e0f21c-fafb-41de-9726-4a840f0a7599-catalog-content\") pod \"redhat-operators-dlpmx\" (UID: \"71e0f21c-fafb-41de-9726-4a840f0a7599\") " pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.759151 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wfqb\" (UniqueName: \"kubernetes.io/projected/71e0f21c-fafb-41de-9726-4a840f0a7599-kube-api-access-6wfqb\") pod \"redhat-operators-dlpmx\" (UID: \"71e0f21c-fafb-41de-9726-4a840f0a7599\") " pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.759797 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71e0f21c-fafb-41de-9726-4a840f0a7599-catalog-content\") pod \"redhat-operators-dlpmx\" (UID: \"71e0f21c-fafb-41de-9726-4a840f0a7599\") " pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.759871 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71e0f21c-fafb-41de-9726-4a840f0a7599-utilities\") pod \"redhat-operators-dlpmx\" (UID: \"71e0f21c-fafb-41de-9726-4a840f0a7599\") " pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.760273 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71e0f21c-fafb-41de-9726-4a840f0a7599-utilities\") pod \"redhat-operators-dlpmx\" (UID: \"71e0f21c-fafb-41de-9726-4a840f0a7599\") " pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.796893 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wfqb\" (UniqueName: \"kubernetes.io/projected/71e0f21c-fafb-41de-9726-4a840f0a7599-kube-api-access-6wfqb\") pod \"redhat-operators-dlpmx\" (UID: \"71e0f21c-fafb-41de-9726-4a840f0a7599\") " pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:23 crc kubenswrapper[4983]: I0318 00:24:23.929481 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:24 crc kubenswrapper[4983]: I0318 00:24:24.226597 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dlpmx"] Mar 18 00:24:25 crc kubenswrapper[4983]: I0318 00:24:25.179156 4983 generic.go:334] "Generic (PLEG): container finished" podID="71e0f21c-fafb-41de-9726-4a840f0a7599" containerID="7b061fbd73b02e8a8823fbabf1c2bc7b402ffd37bfb198c2b7f735c10a6222ff" exitCode=0 Mar 18 00:24:25 crc kubenswrapper[4983]: I0318 00:24:25.179237 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dlpmx" event={"ID":"71e0f21c-fafb-41de-9726-4a840f0a7599","Type":"ContainerDied","Data":"7b061fbd73b02e8a8823fbabf1c2bc7b402ffd37bfb198c2b7f735c10a6222ff"} Mar 18 00:24:25 crc kubenswrapper[4983]: I0318 00:24:25.179499 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dlpmx" event={"ID":"71e0f21c-fafb-41de-9726-4a840f0a7599","Type":"ContainerStarted","Data":"a24b5f49b038eaabc6974b1b7d220b0822d08135f5b6e91bcdb7f8531f0ae995"} Mar 18 00:24:26 crc kubenswrapper[4983]: I0318 00:24:26.187962 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dlpmx" event={"ID":"71e0f21c-fafb-41de-9726-4a840f0a7599","Type":"ContainerStarted","Data":"4f5f216654d4a0553ae7ff2110c1ad0723178849442e0ac0aaaef69901e6e210"} Mar 18 00:24:27 crc kubenswrapper[4983]: I0318 00:24:27.207960 4983 generic.go:334] "Generic (PLEG): container finished" podID="71e0f21c-fafb-41de-9726-4a840f0a7599" containerID="4f5f216654d4a0553ae7ff2110c1ad0723178849442e0ac0aaaef69901e6e210" exitCode=0 Mar 18 00:24:27 crc kubenswrapper[4983]: I0318 00:24:27.208019 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dlpmx" event={"ID":"71e0f21c-fafb-41de-9726-4a840f0a7599","Type":"ContainerDied","Data":"4f5f216654d4a0553ae7ff2110c1ad0723178849442e0ac0aaaef69901e6e210"} Mar 18 00:24:28 crc kubenswrapper[4983]: I0318 00:24:28.215878 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dlpmx" event={"ID":"71e0f21c-fafb-41de-9726-4a840f0a7599","Type":"ContainerStarted","Data":"6b5de129ecaec79447554c181c3e9e39d8cf08a73b166916d7dd96bad4daebbc"} Mar 18 00:24:28 crc kubenswrapper[4983]: I0318 00:24:28.243280 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dlpmx" podStartSLOduration=2.611683451 podStartE2EDuration="5.24325763s" podCreationTimestamp="2026-03-18 00:24:23 +0000 UTC" firstStartedPulling="2026-03-18 00:24:25.182733123 +0000 UTC m=+983.080107733" lastFinishedPulling="2026-03-18 00:24:27.814307262 +0000 UTC m=+985.711681912" observedRunningTime="2026-03-18 00:24:28.238054932 +0000 UTC m=+986.135429562" watchObservedRunningTime="2026-03-18 00:24:28.24325763 +0000 UTC m=+986.140632250" Mar 18 00:24:31 crc kubenswrapper[4983]: I0318 00:24:31.245109 4983 generic.go:334] "Generic (PLEG): container finished" podID="493c24a2-c93e-481a-b2c4-61254df38d13" containerID="1bcd4f1537989b9603ac254a672e191249cbd5ec72167fd79a9bb6fc4b18fe69" exitCode=0 Mar 18 00:24:31 crc kubenswrapper[4983]: I0318 00:24:31.245248 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"493c24a2-c93e-481a-b2c4-61254df38d13","Type":"ContainerDied","Data":"1bcd4f1537989b9603ac254a672e191249cbd5ec72167fd79a9bb6fc4b18fe69"} Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.554887 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.688207 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-ca-bundles\") pod \"493c24a2-c93e-481a-b2c4-61254df38d13\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.688277 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-buildworkdir\") pod \"493c24a2-c93e-481a-b2c4-61254df38d13\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.688320 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-proxy-ca-bundles\") pod \"493c24a2-c93e-481a-b2c4-61254df38d13\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.688346 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-build-blob-cache\") pod \"493c24a2-c93e-481a-b2c4-61254df38d13\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.689181 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "493c24a2-c93e-481a-b2c4-61254df38d13" (UID: "493c24a2-c93e-481a-b2c4-61254df38d13"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.689248 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "493c24a2-c93e-481a-b2c4-61254df38d13" (UID: "493c24a2-c93e-481a-b2c4-61254df38d13"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.697143 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/493c24a2-c93e-481a-b2c4-61254df38d13-builder-dockercfg-q2gzr-push\") pod \"493c24a2-c93e-481a-b2c4-61254df38d13\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.697264 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/493c24a2-c93e-481a-b2c4-61254df38d13-buildcachedir\") pod \"493c24a2-c93e-481a-b2c4-61254df38d13\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.697315 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-container-storage-root\") pod \"493c24a2-c93e-481a-b2c4-61254df38d13\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.697359 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-system-configs\") pod \"493c24a2-c93e-481a-b2c4-61254df38d13\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.697373 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/493c24a2-c93e-481a-b2c4-61254df38d13-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "493c24a2-c93e-481a-b2c4-61254df38d13" (UID: "493c24a2-c93e-481a-b2c4-61254df38d13"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.697413 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/493c24a2-c93e-481a-b2c4-61254df38d13-builder-dockercfg-q2gzr-pull\") pod \"493c24a2-c93e-481a-b2c4-61254df38d13\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.697482 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-container-storage-run\") pod \"493c24a2-c93e-481a-b2c4-61254df38d13\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.697534 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/493c24a2-c93e-481a-b2c4-61254df38d13-node-pullsecrets\") pod \"493c24a2-c93e-481a-b2c4-61254df38d13\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.697591 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfn4x\" (UniqueName: \"kubernetes.io/projected/493c24a2-c93e-481a-b2c4-61254df38d13-kube-api-access-sfn4x\") pod \"493c24a2-c93e-481a-b2c4-61254df38d13\" (UID: \"493c24a2-c93e-481a-b2c4-61254df38d13\") " Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.697681 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/493c24a2-c93e-481a-b2c4-61254df38d13-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "493c24a2-c93e-481a-b2c4-61254df38d13" (UID: "493c24a2-c93e-481a-b2c4-61254df38d13"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.697719 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "493c24a2-c93e-481a-b2c4-61254df38d13" (UID: "493c24a2-c93e-481a-b2c4-61254df38d13"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.698156 4983 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/493c24a2-c93e-481a-b2c4-61254df38d13-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.698190 4983 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.698212 4983 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.698229 4983 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/493c24a2-c93e-481a-b2c4-61254df38d13-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.698243 4983 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/493c24a2-c93e-481a-b2c4-61254df38d13-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.698857 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "493c24a2-c93e-481a-b2c4-61254df38d13" (UID: "493c24a2-c93e-481a-b2c4-61254df38d13"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.701806 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/493c24a2-c93e-481a-b2c4-61254df38d13-kube-api-access-sfn4x" (OuterVolumeSpecName: "kube-api-access-sfn4x") pod "493c24a2-c93e-481a-b2c4-61254df38d13" (UID: "493c24a2-c93e-481a-b2c4-61254df38d13"). InnerVolumeSpecName "kube-api-access-sfn4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.705050 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493c24a2-c93e-481a-b2c4-61254df38d13-builder-dockercfg-q2gzr-push" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-push") pod "493c24a2-c93e-481a-b2c4-61254df38d13" (UID: "493c24a2-c93e-481a-b2c4-61254df38d13"). InnerVolumeSpecName "builder-dockercfg-q2gzr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.706789 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493c24a2-c93e-481a-b2c4-61254df38d13-builder-dockercfg-q2gzr-pull" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-pull") pod "493c24a2-c93e-481a-b2c4-61254df38d13" (UID: "493c24a2-c93e-481a-b2c4-61254df38d13"). InnerVolumeSpecName "builder-dockercfg-q2gzr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.724917 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "493c24a2-c93e-481a-b2c4-61254df38d13" (UID: "493c24a2-c93e-481a-b2c4-61254df38d13"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.802867 4983 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.802904 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/493c24a2-c93e-481a-b2c4-61254df38d13-builder-dockercfg-q2gzr-push\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.802916 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/493c24a2-c93e-481a-b2c4-61254df38d13-builder-dockercfg-q2gzr-pull\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.802945 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.802957 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfn4x\" (UniqueName: \"kubernetes.io/projected/493c24a2-c93e-481a-b2c4-61254df38d13-kube-api-access-sfn4x\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:32 crc kubenswrapper[4983]: I0318 00:24:32.908241 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "493c24a2-c93e-481a-b2c4-61254df38d13" (UID: "493c24a2-c93e-481a-b2c4-61254df38d13"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:24:33 crc kubenswrapper[4983]: I0318 00:24:33.005850 4983 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:33 crc kubenswrapper[4983]: I0318 00:24:33.272041 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"493c24a2-c93e-481a-b2c4-61254df38d13","Type":"ContainerDied","Data":"bef5ca5e98567d173e7048df62149130324df97f1e1a252e909ed0bb77f0d4b8"} Mar 18 00:24:33 crc kubenswrapper[4983]: I0318 00:24:33.272094 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bef5ca5e98567d173e7048df62149130324df97f1e1a252e909ed0bb77f0d4b8" Mar 18 00:24:33 crc kubenswrapper[4983]: I0318 00:24:33.272224 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Mar 18 00:24:33 crc kubenswrapper[4983]: I0318 00:24:33.930092 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:33 crc kubenswrapper[4983]: I0318 00:24:33.930172 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:34 crc kubenswrapper[4983]: I0318 00:24:34.796418 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "493c24a2-c93e-481a-b2c4-61254df38d13" (UID: "493c24a2-c93e-481a-b2c4-61254df38d13"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:24:34 crc kubenswrapper[4983]: I0318 00:24:34.832066 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/493c24a2-c93e-481a-b2c4-61254df38d13-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:34 crc kubenswrapper[4983]: I0318 00:24:34.970751 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dlpmx" podUID="71e0f21c-fafb-41de-9726-4a840f0a7599" containerName="registry-server" probeResult="failure" output=< Mar 18 00:24:34 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Mar 18 00:24:34 crc kubenswrapper[4983]: > Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.575061 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 18 00:24:36 crc kubenswrapper[4983]: E0318 00:24:36.575672 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493c24a2-c93e-481a-b2c4-61254df38d13" containerName="docker-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.575688 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="493c24a2-c93e-481a-b2c4-61254df38d13" containerName="docker-build" Mar 18 00:24:36 crc kubenswrapper[4983]: E0318 00:24:36.575704 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493c24a2-c93e-481a-b2c4-61254df38d13" containerName="git-clone" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.575712 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="493c24a2-c93e-481a-b2c4-61254df38d13" containerName="git-clone" Mar 18 00:24:36 crc kubenswrapper[4983]: E0318 00:24:36.575722 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493c24a2-c93e-481a-b2c4-61254df38d13" containerName="manage-dockerfile" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.575731 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="493c24a2-c93e-481a-b2c4-61254df38d13" containerName="manage-dockerfile" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.575856 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="493c24a2-c93e-481a-b2c4-61254df38d13" containerName="docker-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.576662 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.578781 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-sys-config" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.580038 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-ca" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.581284 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-global-ca" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.599607 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.641527 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-q2gzr" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.660060 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.660141 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.660174 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/e752de16-f363-433a-90a0-d9ac16b98fe4-builder-dockercfg-q2gzr-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.660206 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.660302 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.660341 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e752de16-f363-433a-90a0-d9ac16b98fe4-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.660381 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.660440 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e752de16-f363-433a-90a0-d9ac16b98fe4-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.660519 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/e752de16-f363-433a-90a0-d9ac16b98fe4-builder-dockercfg-q2gzr-push\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.660559 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nnfj\" (UniqueName: \"kubernetes.io/projected/e752de16-f363-433a-90a0-d9ac16b98fe4-kube-api-access-9nnfj\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.660595 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.660637 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.761458 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e752de16-f363-433a-90a0-d9ac16b98fe4-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.761513 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.761541 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e752de16-f363-433a-90a0-d9ac16b98fe4-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.761597 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/e752de16-f363-433a-90a0-d9ac16b98fe4-builder-dockercfg-q2gzr-push\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.761682 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e752de16-f363-433a-90a0-d9ac16b98fe4-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.761761 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e752de16-f363-433a-90a0-d9ac16b98fe4-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.761804 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nnfj\" (UniqueName: \"kubernetes.io/projected/e752de16-f363-433a-90a0-d9ac16b98fe4-kube-api-access-9nnfj\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.761870 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.761990 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.762075 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.762177 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.762199 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.762232 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/e752de16-f363-433a-90a0-d9ac16b98fe4-builder-dockercfg-q2gzr-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.762289 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.762389 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.762402 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.762720 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.763032 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.763078 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.763713 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.764204 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.767153 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/e752de16-f363-433a-90a0-d9ac16b98fe4-builder-dockercfg-q2gzr-push\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.772825 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/e752de16-f363-433a-90a0-d9ac16b98fe4-builder-dockercfg-q2gzr-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.779355 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nnfj\" (UniqueName: \"kubernetes.io/projected/e752de16-f363-433a-90a0-d9ac16b98fe4-kube-api-access-9nnfj\") pod \"smart-gateway-operator-1-build\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:36 crc kubenswrapper[4983]: I0318 00:24:36.948058 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:37 crc kubenswrapper[4983]: I0318 00:24:37.463252 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 18 00:24:37 crc kubenswrapper[4983]: W0318 00:24:37.472181 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode752de16_f363_433a_90a0_d9ac16b98fe4.slice/crio-6f5b306fd2b5bbbae8e15ea80f7fd760e4c3bb9c232df5eb08c7e64995aacdba WatchSource:0}: Error finding container 6f5b306fd2b5bbbae8e15ea80f7fd760e4c3bb9c232df5eb08c7e64995aacdba: Status 404 returned error can't find the container with id 6f5b306fd2b5bbbae8e15ea80f7fd760e4c3bb9c232df5eb08c7e64995aacdba Mar 18 00:24:38 crc kubenswrapper[4983]: I0318 00:24:38.312049 4983 generic.go:334] "Generic (PLEG): container finished" podID="e752de16-f363-433a-90a0-d9ac16b98fe4" containerID="0653dba510531b180f1b933cccd46c5277a373e55ff3662ff5bb5cc685838033" exitCode=0 Mar 18 00:24:38 crc kubenswrapper[4983]: I0318 00:24:38.312147 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"e752de16-f363-433a-90a0-d9ac16b98fe4","Type":"ContainerDied","Data":"0653dba510531b180f1b933cccd46c5277a373e55ff3662ff5bb5cc685838033"} Mar 18 00:24:38 crc kubenswrapper[4983]: I0318 00:24:38.313406 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"e752de16-f363-433a-90a0-d9ac16b98fe4","Type":"ContainerStarted","Data":"6f5b306fd2b5bbbae8e15ea80f7fd760e4c3bb9c232df5eb08c7e64995aacdba"} Mar 18 00:24:39 crc kubenswrapper[4983]: I0318 00:24:39.324204 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"e752de16-f363-433a-90a0-d9ac16b98fe4","Type":"ContainerStarted","Data":"59d332cfd7110c1599b6f82fc9817c7a5168d59675a149ea8b1576d43ad06669"} Mar 18 00:24:39 crc kubenswrapper[4983]: I0318 00:24:39.356259 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-1-build" podStartSLOduration=3.356225369 podStartE2EDuration="3.356225369s" podCreationTimestamp="2026-03-18 00:24:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:24:39.351818474 +0000 UTC m=+997.249193174" watchObservedRunningTime="2026-03-18 00:24:39.356225369 +0000 UTC m=+997.253600029" Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.053083 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d2g4j"] Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.054604 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.109465 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d2g4j"] Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.112459 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blmsv\" (UniqueName: \"kubernetes.io/projected/f01fa7bc-915f-4076-a5e8-8bc01bb745af-kube-api-access-blmsv\") pod \"certified-operators-d2g4j\" (UID: \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\") " pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.112503 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f01fa7bc-915f-4076-a5e8-8bc01bb745af-catalog-content\") pod \"certified-operators-d2g4j\" (UID: \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\") " pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.112567 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f01fa7bc-915f-4076-a5e8-8bc01bb745af-utilities\") pod \"certified-operators-d2g4j\" (UID: \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\") " pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.213811 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blmsv\" (UniqueName: \"kubernetes.io/projected/f01fa7bc-915f-4076-a5e8-8bc01bb745af-kube-api-access-blmsv\") pod \"certified-operators-d2g4j\" (UID: \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\") " pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.213868 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f01fa7bc-915f-4076-a5e8-8bc01bb745af-catalog-content\") pod \"certified-operators-d2g4j\" (UID: \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\") " pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.213908 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f01fa7bc-915f-4076-a5e8-8bc01bb745af-utilities\") pod \"certified-operators-d2g4j\" (UID: \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\") " pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.214371 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f01fa7bc-915f-4076-a5e8-8bc01bb745af-utilities\") pod \"certified-operators-d2g4j\" (UID: \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\") " pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.214457 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f01fa7bc-915f-4076-a5e8-8bc01bb745af-catalog-content\") pod \"certified-operators-d2g4j\" (UID: \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\") " pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.231471 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blmsv\" (UniqueName: \"kubernetes.io/projected/f01fa7bc-915f-4076-a5e8-8bc01bb745af-kube-api-access-blmsv\") pod \"certified-operators-d2g4j\" (UID: \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\") " pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.418973 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:40 crc kubenswrapper[4983]: I0318 00:24:40.658633 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d2g4j"] Mar 18 00:24:41 crc kubenswrapper[4983]: I0318 00:24:41.344423 4983 generic.go:334] "Generic (PLEG): container finished" podID="f01fa7bc-915f-4076-a5e8-8bc01bb745af" containerID="4f6520530aa53a878471e4e2687293b15e28ea02fc847a570e9de1f19c8f9252" exitCode=0 Mar 18 00:24:41 crc kubenswrapper[4983]: I0318 00:24:41.344747 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2g4j" event={"ID":"f01fa7bc-915f-4076-a5e8-8bc01bb745af","Type":"ContainerDied","Data":"4f6520530aa53a878471e4e2687293b15e28ea02fc847a570e9de1f19c8f9252"} Mar 18 00:24:41 crc kubenswrapper[4983]: I0318 00:24:41.344783 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2g4j" event={"ID":"f01fa7bc-915f-4076-a5e8-8bc01bb745af","Type":"ContainerStarted","Data":"268d822a18f1e12eba9e60a6f593595c8fdf4d9b068a86c9162fc01a38e1a2bf"} Mar 18 00:24:42 crc kubenswrapper[4983]: I0318 00:24:42.356556 4983 generic.go:334] "Generic (PLEG): container finished" podID="f01fa7bc-915f-4076-a5e8-8bc01bb745af" containerID="65f68d3e8aeb57fde85764cf459ca9da645245cc3bcb1dd9e27c8af9e88e0c3e" exitCode=0 Mar 18 00:24:42 crc kubenswrapper[4983]: I0318 00:24:42.356602 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2g4j" event={"ID":"f01fa7bc-915f-4076-a5e8-8bc01bb745af","Type":"ContainerDied","Data":"65f68d3e8aeb57fde85764cf459ca9da645245cc3bcb1dd9e27c8af9e88e0c3e"} Mar 18 00:24:43 crc kubenswrapper[4983]: I0318 00:24:43.258970 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:24:43 crc kubenswrapper[4983]: I0318 00:24:43.259311 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:24:43 crc kubenswrapper[4983]: I0318 00:24:43.259358 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:24:43 crc kubenswrapper[4983]: I0318 00:24:43.259808 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"81d24af741ccda81f14aaade4f3fc2f7c8569f253f97a8c8f9b6fa245fbd1638"} pod="openshift-machine-config-operator/machine-config-daemon-jczfs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 00:24:43 crc kubenswrapper[4983]: I0318 00:24:43.259863 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" containerID="cri-o://81d24af741ccda81f14aaade4f3fc2f7c8569f253f97a8c8f9b6fa245fbd1638" gracePeriod=600 Mar 18 00:24:43 crc kubenswrapper[4983]: I0318 00:24:43.366161 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2g4j" event={"ID":"f01fa7bc-915f-4076-a5e8-8bc01bb745af","Type":"ContainerStarted","Data":"39212d0b4bdff74a3a80cc7b73029334c760ccdbe73ce7e80ea61acba086b8e0"} Mar 18 00:24:43 crc kubenswrapper[4983]: I0318 00:24:43.394087 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d2g4j" podStartSLOduration=1.979313187 podStartE2EDuration="3.394068709s" podCreationTimestamp="2026-03-18 00:24:40 +0000 UTC" firstStartedPulling="2026-03-18 00:24:41.347361151 +0000 UTC m=+999.244735761" lastFinishedPulling="2026-03-18 00:24:42.762116673 +0000 UTC m=+1000.659491283" observedRunningTime="2026-03-18 00:24:43.391311091 +0000 UTC m=+1001.288685731" watchObservedRunningTime="2026-03-18 00:24:43.394068709 +0000 UTC m=+1001.291443319" Mar 18 00:24:43 crc kubenswrapper[4983]: I0318 00:24:43.991159 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:44 crc kubenswrapper[4983]: I0318 00:24:44.033192 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:44 crc kubenswrapper[4983]: I0318 00:24:44.383353 4983 generic.go:334] "Generic (PLEG): container finished" podID="482e2afe-48be-4ea2-8bda-52849c901b52" containerID="81d24af741ccda81f14aaade4f3fc2f7c8569f253f97a8c8f9b6fa245fbd1638" exitCode=0 Mar 18 00:24:44 crc kubenswrapper[4983]: I0318 00:24:44.383448 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerDied","Data":"81d24af741ccda81f14aaade4f3fc2f7c8569f253f97a8c8f9b6fa245fbd1638"} Mar 18 00:24:44 crc kubenswrapper[4983]: I0318 00:24:44.383887 4983 scope.go:117] "RemoveContainer" containerID="9cc35c9c62ec6606ad35039f3a8a4340518e27795998f82a00d733ef5cac33bb" Mar 18 00:24:44 crc kubenswrapper[4983]: I0318 00:24:44.384451 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerStarted","Data":"f0dbf5791681286cf38e89c3a6323146d81246c217b7c32c7cf2e180e155fbe8"} Mar 18 00:24:46 crc kubenswrapper[4983]: I0318 00:24:46.236222 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dlpmx"] Mar 18 00:24:46 crc kubenswrapper[4983]: I0318 00:24:46.236767 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dlpmx" podUID="71e0f21c-fafb-41de-9726-4a840f0a7599" containerName="registry-server" containerID="cri-o://6b5de129ecaec79447554c181c3e9e39d8cf08a73b166916d7dd96bad4daebbc" gracePeriod=2 Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.155732 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.156175 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/smart-gateway-operator-1-build" podUID="e752de16-f363-433a-90a0-d9ac16b98fe4" containerName="docker-build" containerID="cri-o://59d332cfd7110c1599b6f82fc9817c7a5168d59675a149ea8b1576d43ad06669" gracePeriod=30 Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.228866 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.321788 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71e0f21c-fafb-41de-9726-4a840f0a7599-utilities\") pod \"71e0f21c-fafb-41de-9726-4a840f0a7599\" (UID: \"71e0f21c-fafb-41de-9726-4a840f0a7599\") " Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.321960 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71e0f21c-fafb-41de-9726-4a840f0a7599-catalog-content\") pod \"71e0f21c-fafb-41de-9726-4a840f0a7599\" (UID: \"71e0f21c-fafb-41de-9726-4a840f0a7599\") " Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.322194 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wfqb\" (UniqueName: \"kubernetes.io/projected/71e0f21c-fafb-41de-9726-4a840f0a7599-kube-api-access-6wfqb\") pod \"71e0f21c-fafb-41de-9726-4a840f0a7599\" (UID: \"71e0f21c-fafb-41de-9726-4a840f0a7599\") " Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.323333 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71e0f21c-fafb-41de-9726-4a840f0a7599-utilities" (OuterVolumeSpecName: "utilities") pod "71e0f21c-fafb-41de-9726-4a840f0a7599" (UID: "71e0f21c-fafb-41de-9726-4a840f0a7599"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.331424 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71e0f21c-fafb-41de-9726-4a840f0a7599-kube-api-access-6wfqb" (OuterVolumeSpecName: "kube-api-access-6wfqb") pod "71e0f21c-fafb-41de-9726-4a840f0a7599" (UID: "71e0f21c-fafb-41de-9726-4a840f0a7599"). InnerVolumeSpecName "kube-api-access-6wfqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.419857 4983 generic.go:334] "Generic (PLEG): container finished" podID="71e0f21c-fafb-41de-9726-4a840f0a7599" containerID="6b5de129ecaec79447554c181c3e9e39d8cf08a73b166916d7dd96bad4daebbc" exitCode=0 Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.419948 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dlpmx" event={"ID":"71e0f21c-fafb-41de-9726-4a840f0a7599","Type":"ContainerDied","Data":"6b5de129ecaec79447554c181c3e9e39d8cf08a73b166916d7dd96bad4daebbc"} Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.419980 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dlpmx" event={"ID":"71e0f21c-fafb-41de-9726-4a840f0a7599","Type":"ContainerDied","Data":"a24b5f49b038eaabc6974b1b7d220b0822d08135f5b6e91bcdb7f8531f0ae995"} Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.420001 4983 scope.go:117] "RemoveContainer" containerID="6b5de129ecaec79447554c181c3e9e39d8cf08a73b166916d7dd96bad4daebbc" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.420207 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dlpmx" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.427825 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wfqb\" (UniqueName: \"kubernetes.io/projected/71e0f21c-fafb-41de-9726-4a840f0a7599-kube-api-access-6wfqb\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.427863 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71e0f21c-fafb-41de-9726-4a840f0a7599-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.439579 4983 scope.go:117] "RemoveContainer" containerID="4f5f216654d4a0553ae7ff2110c1ad0723178849442e0ac0aaaef69901e6e210" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.449014 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71e0f21c-fafb-41de-9726-4a840f0a7599-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71e0f21c-fafb-41de-9726-4a840f0a7599" (UID: "71e0f21c-fafb-41de-9726-4a840f0a7599"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.461342 4983 scope.go:117] "RemoveContainer" containerID="7b061fbd73b02e8a8823fbabf1c2bc7b402ffd37bfb198c2b7f735c10a6222ff" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.488985 4983 scope.go:117] "RemoveContainer" containerID="6b5de129ecaec79447554c181c3e9e39d8cf08a73b166916d7dd96bad4daebbc" Mar 18 00:24:47 crc kubenswrapper[4983]: E0318 00:24:47.489396 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b5de129ecaec79447554c181c3e9e39d8cf08a73b166916d7dd96bad4daebbc\": container with ID starting with 6b5de129ecaec79447554c181c3e9e39d8cf08a73b166916d7dd96bad4daebbc not found: ID does not exist" containerID="6b5de129ecaec79447554c181c3e9e39d8cf08a73b166916d7dd96bad4daebbc" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.489457 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b5de129ecaec79447554c181c3e9e39d8cf08a73b166916d7dd96bad4daebbc"} err="failed to get container status \"6b5de129ecaec79447554c181c3e9e39d8cf08a73b166916d7dd96bad4daebbc\": rpc error: code = NotFound desc = could not find container \"6b5de129ecaec79447554c181c3e9e39d8cf08a73b166916d7dd96bad4daebbc\": container with ID starting with 6b5de129ecaec79447554c181c3e9e39d8cf08a73b166916d7dd96bad4daebbc not found: ID does not exist" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.489490 4983 scope.go:117] "RemoveContainer" containerID="4f5f216654d4a0553ae7ff2110c1ad0723178849442e0ac0aaaef69901e6e210" Mar 18 00:24:47 crc kubenswrapper[4983]: E0318 00:24:47.489839 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f5f216654d4a0553ae7ff2110c1ad0723178849442e0ac0aaaef69901e6e210\": container with ID starting with 4f5f216654d4a0553ae7ff2110c1ad0723178849442e0ac0aaaef69901e6e210 not found: ID does not exist" containerID="4f5f216654d4a0553ae7ff2110c1ad0723178849442e0ac0aaaef69901e6e210" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.489874 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f5f216654d4a0553ae7ff2110c1ad0723178849442e0ac0aaaef69901e6e210"} err="failed to get container status \"4f5f216654d4a0553ae7ff2110c1ad0723178849442e0ac0aaaef69901e6e210\": rpc error: code = NotFound desc = could not find container \"4f5f216654d4a0553ae7ff2110c1ad0723178849442e0ac0aaaef69901e6e210\": container with ID starting with 4f5f216654d4a0553ae7ff2110c1ad0723178849442e0ac0aaaef69901e6e210 not found: ID does not exist" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.489913 4983 scope.go:117] "RemoveContainer" containerID="7b061fbd73b02e8a8823fbabf1c2bc7b402ffd37bfb198c2b7f735c10a6222ff" Mar 18 00:24:47 crc kubenswrapper[4983]: E0318 00:24:47.490423 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b061fbd73b02e8a8823fbabf1c2bc7b402ffd37bfb198c2b7f735c10a6222ff\": container with ID starting with 7b061fbd73b02e8a8823fbabf1c2bc7b402ffd37bfb198c2b7f735c10a6222ff not found: ID does not exist" containerID="7b061fbd73b02e8a8823fbabf1c2bc7b402ffd37bfb198c2b7f735c10a6222ff" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.490477 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b061fbd73b02e8a8823fbabf1c2bc7b402ffd37bfb198c2b7f735c10a6222ff"} err="failed to get container status \"7b061fbd73b02e8a8823fbabf1c2bc7b402ffd37bfb198c2b7f735c10a6222ff\": rpc error: code = NotFound desc = could not find container \"7b061fbd73b02e8a8823fbabf1c2bc7b402ffd37bfb198c2b7f735c10a6222ff\": container with ID starting with 7b061fbd73b02e8a8823fbabf1c2bc7b402ffd37bfb198c2b7f735c10a6222ff not found: ID does not exist" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.529633 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71e0f21c-fafb-41de-9726-4a840f0a7599-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.743101 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dlpmx"] Mar 18 00:24:47 crc kubenswrapper[4983]: I0318 00:24:47.748914 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dlpmx"] Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.830332 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Mar 18 00:24:48 crc kubenswrapper[4983]: E0318 00:24:48.833177 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e0f21c-fafb-41de-9726-4a840f0a7599" containerName="extract-utilities" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.833387 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e0f21c-fafb-41de-9726-4a840f0a7599" containerName="extract-utilities" Mar 18 00:24:48 crc kubenswrapper[4983]: E0318 00:24:48.833518 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e0f21c-fafb-41de-9726-4a840f0a7599" containerName="extract-content" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.833635 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e0f21c-fafb-41de-9726-4a840f0a7599" containerName="extract-content" Mar 18 00:24:48 crc kubenswrapper[4983]: E0318 00:24:48.833839 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e0f21c-fafb-41de-9726-4a840f0a7599" containerName="registry-server" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.834045 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e0f21c-fafb-41de-9726-4a840f0a7599" containerName="registry-server" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.834453 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="71e0f21c-fafb-41de-9726-4a840f0a7599" containerName="registry-server" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.836572 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.839846 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-sys-config" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.840270 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-ca" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.840477 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-global-ca" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.848382 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqgr2\" (UniqueName: \"kubernetes.io/projected/a507119f-75ce-48dc-ba8d-374e9c39cd09-kube-api-access-wqgr2\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.848523 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.848584 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a507119f-75ce-48dc-ba8d-374e9c39cd09-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.848665 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.848765 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/a507119f-75ce-48dc-ba8d-374e9c39cd09-builder-dockercfg-q2gzr-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.848826 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.848877 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.848925 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a507119f-75ce-48dc-ba8d-374e9c39cd09-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.848990 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/a507119f-75ce-48dc-ba8d-374e9c39cd09-builder-dockercfg-q2gzr-push\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.849031 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.849070 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.849115 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.896173 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71e0f21c-fafb-41de-9726-4a840f0a7599" path="/var/lib/kubelet/pods/71e0f21c-fafb-41de-9726-4a840f0a7599/volumes" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.897007 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.950421 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.950499 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a507119f-75ce-48dc-ba8d-374e9c39cd09-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.950525 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.950573 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/a507119f-75ce-48dc-ba8d-374e9c39cd09-builder-dockercfg-q2gzr-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.950603 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.950627 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.950650 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a507119f-75ce-48dc-ba8d-374e9c39cd09-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.950673 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.950691 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/a507119f-75ce-48dc-ba8d-374e9c39cd09-builder-dockercfg-q2gzr-push\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.950709 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.950732 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.950774 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqgr2\" (UniqueName: \"kubernetes.io/projected/a507119f-75ce-48dc-ba8d-374e9c39cd09-kube-api-access-wqgr2\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.951836 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.951901 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a507119f-75ce-48dc-ba8d-374e9c39cd09-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.952162 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.953134 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.953159 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a507119f-75ce-48dc-ba8d-374e9c39cd09-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.953486 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.954244 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.954571 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.954702 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.958714 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/a507119f-75ce-48dc-ba8d-374e9c39cd09-builder-dockercfg-q2gzr-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.961663 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/a507119f-75ce-48dc-ba8d-374e9c39cd09-builder-dockercfg-q2gzr-push\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:48 crc kubenswrapper[4983]: I0318 00:24:48.977408 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqgr2\" (UniqueName: \"kubernetes.io/projected/a507119f-75ce-48dc-ba8d-374e9c39cd09-kube-api-access-wqgr2\") pod \"smart-gateway-operator-2-build\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:49 crc kubenswrapper[4983]: I0318 00:24:49.167588 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:24:49 crc kubenswrapper[4983]: I0318 00:24:49.414619 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Mar 18 00:24:49 crc kubenswrapper[4983]: W0318 00:24:49.425855 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda507119f_75ce_48dc_ba8d_374e9c39cd09.slice/crio-c2a3b3a058c50f42dc7b9c24a1590b42694128f5fa1053a6b122de4c19db241e WatchSource:0}: Error finding container c2a3b3a058c50f42dc7b9c24a1590b42694128f5fa1053a6b122de4c19db241e: Status 404 returned error can't find the container with id c2a3b3a058c50f42dc7b9c24a1590b42694128f5fa1053a6b122de4c19db241e Mar 18 00:24:49 crc kubenswrapper[4983]: I0318 00:24:49.438998 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a507119f-75ce-48dc-ba8d-374e9c39cd09","Type":"ContainerStarted","Data":"c2a3b3a058c50f42dc7b9c24a1590b42694128f5fa1053a6b122de4c19db241e"} Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.296446 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_e752de16-f363-433a-90a0-d9ac16b98fe4/docker-build/0.log" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.301434 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.371294 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/e752de16-f363-433a-90a0-d9ac16b98fe4-builder-dockercfg-q2gzr-push\") pod \"e752de16-f363-433a-90a0-d9ac16b98fe4\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.371397 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e752de16-f363-433a-90a0-d9ac16b98fe4-buildcachedir\") pod \"e752de16-f363-433a-90a0-d9ac16b98fe4\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.371476 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e752de16-f363-433a-90a0-d9ac16b98fe4-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "e752de16-f363-433a-90a0-d9ac16b98fe4" (UID: "e752de16-f363-433a-90a0-d9ac16b98fe4"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.371518 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-container-storage-root\") pod \"e752de16-f363-433a-90a0-d9ac16b98fe4\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.371545 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nnfj\" (UniqueName: \"kubernetes.io/projected/e752de16-f363-433a-90a0-d9ac16b98fe4-kube-api-access-9nnfj\") pod \"e752de16-f363-433a-90a0-d9ac16b98fe4\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.372035 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "e752de16-f363-433a-90a0-d9ac16b98fe4" (UID: "e752de16-f363-433a-90a0-d9ac16b98fe4"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.372328 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-buildworkdir\") pod \"e752de16-f363-433a-90a0-d9ac16b98fe4\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.372389 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-ca-bundles\") pod \"e752de16-f363-433a-90a0-d9ac16b98fe4\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.372422 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-build-blob-cache\") pod \"e752de16-f363-433a-90a0-d9ac16b98fe4\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.372460 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-container-storage-run\") pod \"e752de16-f363-433a-90a0-d9ac16b98fe4\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.372491 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e752de16-f363-433a-90a0-d9ac16b98fe4-node-pullsecrets\") pod \"e752de16-f363-433a-90a0-d9ac16b98fe4\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.372529 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-system-configs\") pod \"e752de16-f363-433a-90a0-d9ac16b98fe4\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.372561 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-proxy-ca-bundles\") pod \"e752de16-f363-433a-90a0-d9ac16b98fe4\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.372585 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/e752de16-f363-433a-90a0-d9ac16b98fe4-builder-dockercfg-q2gzr-pull\") pod \"e752de16-f363-433a-90a0-d9ac16b98fe4\" (UID: \"e752de16-f363-433a-90a0-d9ac16b98fe4\") " Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.372960 4983 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e752de16-f363-433a-90a0-d9ac16b98fe4-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.372983 4983 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.373293 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "e752de16-f363-433a-90a0-d9ac16b98fe4" (UID: "e752de16-f363-433a-90a0-d9ac16b98fe4"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.373617 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "e752de16-f363-433a-90a0-d9ac16b98fe4" (UID: "e752de16-f363-433a-90a0-d9ac16b98fe4"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.374286 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "e752de16-f363-433a-90a0-d9ac16b98fe4" (UID: "e752de16-f363-433a-90a0-d9ac16b98fe4"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.374334 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e752de16-f363-433a-90a0-d9ac16b98fe4-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "e752de16-f363-433a-90a0-d9ac16b98fe4" (UID: "e752de16-f363-433a-90a0-d9ac16b98fe4"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.374585 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "e752de16-f363-433a-90a0-d9ac16b98fe4" (UID: "e752de16-f363-433a-90a0-d9ac16b98fe4"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.378192 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e752de16-f363-433a-90a0-d9ac16b98fe4-kube-api-access-9nnfj" (OuterVolumeSpecName: "kube-api-access-9nnfj") pod "e752de16-f363-433a-90a0-d9ac16b98fe4" (UID: "e752de16-f363-433a-90a0-d9ac16b98fe4"). InnerVolumeSpecName "kube-api-access-9nnfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.378419 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e752de16-f363-433a-90a0-d9ac16b98fe4-builder-dockercfg-q2gzr-pull" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-pull") pod "e752de16-f363-433a-90a0-d9ac16b98fe4" (UID: "e752de16-f363-433a-90a0-d9ac16b98fe4"). InnerVolumeSpecName "builder-dockercfg-q2gzr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.378516 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e752de16-f363-433a-90a0-d9ac16b98fe4-builder-dockercfg-q2gzr-push" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-push") pod "e752de16-f363-433a-90a0-d9ac16b98fe4" (UID: "e752de16-f363-433a-90a0-d9ac16b98fe4"). InnerVolumeSpecName "builder-dockercfg-q2gzr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.419317 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.419367 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.446551 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_e752de16-f363-433a-90a0-d9ac16b98fe4/docker-build/0.log" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.447009 4983 generic.go:334] "Generic (PLEG): container finished" podID="e752de16-f363-433a-90a0-d9ac16b98fe4" containerID="59d332cfd7110c1599b6f82fc9817c7a5168d59675a149ea8b1576d43ad06669" exitCode=1 Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.447211 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"e752de16-f363-433a-90a0-d9ac16b98fe4","Type":"ContainerDied","Data":"59d332cfd7110c1599b6f82fc9817c7a5168d59675a149ea8b1576d43ad06669"} Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.447286 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"e752de16-f363-433a-90a0-d9ac16b98fe4","Type":"ContainerDied","Data":"6f5b306fd2b5bbbae8e15ea80f7fd760e4c3bb9c232df5eb08c7e64995aacdba"} Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.447309 4983 scope.go:117] "RemoveContainer" containerID="59d332cfd7110c1599b6f82fc9817c7a5168d59675a149ea8b1576d43ad06669" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.447553 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.454402 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a507119f-75ce-48dc-ba8d-374e9c39cd09","Type":"ContainerStarted","Data":"049bc754ae68db77841ac7269856b798fe2439f8b885d8dbb233a1bad34680db"} Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.459509 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.486419 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/e752de16-f363-433a-90a0-d9ac16b98fe4-builder-dockercfg-q2gzr-push\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.486446 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nnfj\" (UniqueName: \"kubernetes.io/projected/e752de16-f363-433a-90a0-d9ac16b98fe4-kube-api-access-9nnfj\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.486456 4983 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.486465 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.486474 4983 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e752de16-f363-433a-90a0-d9ac16b98fe4-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.486482 4983 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.486490 4983 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e752de16-f363-433a-90a0-d9ac16b98fe4-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.486517 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/e752de16-f363-433a-90a0-d9ac16b98fe4-builder-dockercfg-q2gzr-pull\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.502572 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.514875 4983 scope.go:117] "RemoveContainer" containerID="0653dba510531b180f1b933cccd46c5277a373e55ff3662ff5bb5cc685838033" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.519871 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "e752de16-f363-433a-90a0-d9ac16b98fe4" (UID: "e752de16-f363-433a-90a0-d9ac16b98fe4"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.555105 4983 scope.go:117] "RemoveContainer" containerID="59d332cfd7110c1599b6f82fc9817c7a5168d59675a149ea8b1576d43ad06669" Mar 18 00:24:50 crc kubenswrapper[4983]: E0318 00:24:50.555667 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59d332cfd7110c1599b6f82fc9817c7a5168d59675a149ea8b1576d43ad06669\": container with ID starting with 59d332cfd7110c1599b6f82fc9817c7a5168d59675a149ea8b1576d43ad06669 not found: ID does not exist" containerID="59d332cfd7110c1599b6f82fc9817c7a5168d59675a149ea8b1576d43ad06669" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.555711 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59d332cfd7110c1599b6f82fc9817c7a5168d59675a149ea8b1576d43ad06669"} err="failed to get container status \"59d332cfd7110c1599b6f82fc9817c7a5168d59675a149ea8b1576d43ad06669\": rpc error: code = NotFound desc = could not find container \"59d332cfd7110c1599b6f82fc9817c7a5168d59675a149ea8b1576d43ad06669\": container with ID starting with 59d332cfd7110c1599b6f82fc9817c7a5168d59675a149ea8b1576d43ad06669 not found: ID does not exist" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.555742 4983 scope.go:117] "RemoveContainer" containerID="0653dba510531b180f1b933cccd46c5277a373e55ff3662ff5bb5cc685838033" Mar 18 00:24:50 crc kubenswrapper[4983]: E0318 00:24:50.556426 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0653dba510531b180f1b933cccd46c5277a373e55ff3662ff5bb5cc685838033\": container with ID starting with 0653dba510531b180f1b933cccd46c5277a373e55ff3662ff5bb5cc685838033 not found: ID does not exist" containerID="0653dba510531b180f1b933cccd46c5277a373e55ff3662ff5bb5cc685838033" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.556461 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0653dba510531b180f1b933cccd46c5277a373e55ff3662ff5bb5cc685838033"} err="failed to get container status \"0653dba510531b180f1b933cccd46c5277a373e55ff3662ff5bb5cc685838033\": rpc error: code = NotFound desc = could not find container \"0653dba510531b180f1b933cccd46c5277a373e55ff3662ff5bb5cc685838033\": container with ID starting with 0653dba510531b180f1b933cccd46c5277a373e55ff3662ff5bb5cc685838033 not found: ID does not exist" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.587521 4983 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.761919 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "e752de16-f363-433a-90a0-d9ac16b98fe4" (UID: "e752de16-f363-433a-90a0-d9ac16b98fe4"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:24:50 crc kubenswrapper[4983]: I0318 00:24:50.790036 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e752de16-f363-433a-90a0-d9ac16b98fe4-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:51 crc kubenswrapper[4983]: I0318 00:24:51.074476 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 18 00:24:51 crc kubenswrapper[4983]: I0318 00:24:51.086134 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 18 00:24:51 crc kubenswrapper[4983]: I0318 00:24:51.469600 4983 generic.go:334] "Generic (PLEG): container finished" podID="a507119f-75ce-48dc-ba8d-374e9c39cd09" containerID="049bc754ae68db77841ac7269856b798fe2439f8b885d8dbb233a1bad34680db" exitCode=0 Mar 18 00:24:51 crc kubenswrapper[4983]: I0318 00:24:51.470610 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a507119f-75ce-48dc-ba8d-374e9c39cd09","Type":"ContainerDied","Data":"049bc754ae68db77841ac7269856b798fe2439f8b885d8dbb233a1bad34680db"} Mar 18 00:24:51 crc kubenswrapper[4983]: I0318 00:24:51.639662 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d2g4j"] Mar 18 00:24:52 crc kubenswrapper[4983]: I0318 00:24:52.487089 4983 generic.go:334] "Generic (PLEG): container finished" podID="a507119f-75ce-48dc-ba8d-374e9c39cd09" containerID="8b91275e7db5894979dddab3895e7f591cc71361a914a4b71320ab58d11498a9" exitCode=0 Mar 18 00:24:52 crc kubenswrapper[4983]: I0318 00:24:52.487747 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d2g4j" podUID="f01fa7bc-915f-4076-a5e8-8bc01bb745af" containerName="registry-server" containerID="cri-o://39212d0b4bdff74a3a80cc7b73029334c760ccdbe73ce7e80ea61acba086b8e0" gracePeriod=2 Mar 18 00:24:52 crc kubenswrapper[4983]: I0318 00:24:52.487240 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a507119f-75ce-48dc-ba8d-374e9c39cd09","Type":"ContainerDied","Data":"8b91275e7db5894979dddab3895e7f591cc71361a914a4b71320ab58d11498a9"} Mar 18 00:24:52 crc kubenswrapper[4983]: I0318 00:24:52.551198 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-2-build_a507119f-75ce-48dc-ba8d-374e9c39cd09/manage-dockerfile/0.log" Mar 18 00:24:52 crc kubenswrapper[4983]: I0318 00:24:52.854762 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e752de16-f363-433a-90a0-d9ac16b98fe4" path="/var/lib/kubelet/pods/e752de16-f363-433a-90a0-d9ac16b98fe4/volumes" Mar 18 00:24:52 crc kubenswrapper[4983]: I0318 00:24:52.886718 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:52 crc kubenswrapper[4983]: I0318 00:24:52.919279 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f01fa7bc-915f-4076-a5e8-8bc01bb745af-utilities\") pod \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\" (UID: \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\") " Mar 18 00:24:52 crc kubenswrapper[4983]: I0318 00:24:52.919383 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f01fa7bc-915f-4076-a5e8-8bc01bb745af-catalog-content\") pod \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\" (UID: \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\") " Mar 18 00:24:52 crc kubenswrapper[4983]: I0318 00:24:52.919421 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blmsv\" (UniqueName: \"kubernetes.io/projected/f01fa7bc-915f-4076-a5e8-8bc01bb745af-kube-api-access-blmsv\") pod \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\" (UID: \"f01fa7bc-915f-4076-a5e8-8bc01bb745af\") " Mar 18 00:24:52 crc kubenswrapper[4983]: I0318 00:24:52.924707 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f01fa7bc-915f-4076-a5e8-8bc01bb745af-kube-api-access-blmsv" (OuterVolumeSpecName: "kube-api-access-blmsv") pod "f01fa7bc-915f-4076-a5e8-8bc01bb745af" (UID: "f01fa7bc-915f-4076-a5e8-8bc01bb745af"). InnerVolumeSpecName "kube-api-access-blmsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:24:52 crc kubenswrapper[4983]: I0318 00:24:52.926628 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f01fa7bc-915f-4076-a5e8-8bc01bb745af-utilities" (OuterVolumeSpecName: "utilities") pod "f01fa7bc-915f-4076-a5e8-8bc01bb745af" (UID: "f01fa7bc-915f-4076-a5e8-8bc01bb745af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:24:52 crc kubenswrapper[4983]: I0318 00:24:52.971576 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f01fa7bc-915f-4076-a5e8-8bc01bb745af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f01fa7bc-915f-4076-a5e8-8bc01bb745af" (UID: "f01fa7bc-915f-4076-a5e8-8bc01bb745af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.020791 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blmsv\" (UniqueName: \"kubernetes.io/projected/f01fa7bc-915f-4076-a5e8-8bc01bb745af-kube-api-access-blmsv\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.020825 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f01fa7bc-915f-4076-a5e8-8bc01bb745af-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.020838 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f01fa7bc-915f-4076-a5e8-8bc01bb745af-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.511222 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a507119f-75ce-48dc-ba8d-374e9c39cd09","Type":"ContainerStarted","Data":"5636a5b2375cfdbaae981b745551fae086f4c3a9b4777b501f5f8a4649de5d3e"} Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.515371 4983 generic.go:334] "Generic (PLEG): container finished" podID="f01fa7bc-915f-4076-a5e8-8bc01bb745af" containerID="39212d0b4bdff74a3a80cc7b73029334c760ccdbe73ce7e80ea61acba086b8e0" exitCode=0 Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.515450 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2g4j" event={"ID":"f01fa7bc-915f-4076-a5e8-8bc01bb745af","Type":"ContainerDied","Data":"39212d0b4bdff74a3a80cc7b73029334c760ccdbe73ce7e80ea61acba086b8e0"} Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.515473 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d2g4j" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.515522 4983 scope.go:117] "RemoveContainer" containerID="39212d0b4bdff74a3a80cc7b73029334c760ccdbe73ce7e80ea61acba086b8e0" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.515502 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d2g4j" event={"ID":"f01fa7bc-915f-4076-a5e8-8bc01bb745af","Type":"ContainerDied","Data":"268d822a18f1e12eba9e60a6f593595c8fdf4d9b068a86c9162fc01a38e1a2bf"} Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.545556 4983 scope.go:117] "RemoveContainer" containerID="65f68d3e8aeb57fde85764cf459ca9da645245cc3bcb1dd9e27c8af9e88e0c3e" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.571179 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-2-build" podStartSLOduration=5.57115152 podStartE2EDuration="5.57115152s" podCreationTimestamp="2026-03-18 00:24:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:24:53.565884421 +0000 UTC m=+1011.463259071" watchObservedRunningTime="2026-03-18 00:24:53.57115152 +0000 UTC m=+1011.468526160" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.594404 4983 scope.go:117] "RemoveContainer" containerID="4f6520530aa53a878471e4e2687293b15e28ea02fc847a570e9de1f19c8f9252" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.598485 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d2g4j"] Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.607433 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d2g4j"] Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.615157 4983 scope.go:117] "RemoveContainer" containerID="39212d0b4bdff74a3a80cc7b73029334c760ccdbe73ce7e80ea61acba086b8e0" Mar 18 00:24:53 crc kubenswrapper[4983]: E0318 00:24:53.616074 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39212d0b4bdff74a3a80cc7b73029334c760ccdbe73ce7e80ea61acba086b8e0\": container with ID starting with 39212d0b4bdff74a3a80cc7b73029334c760ccdbe73ce7e80ea61acba086b8e0 not found: ID does not exist" containerID="39212d0b4bdff74a3a80cc7b73029334c760ccdbe73ce7e80ea61acba086b8e0" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.616139 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39212d0b4bdff74a3a80cc7b73029334c760ccdbe73ce7e80ea61acba086b8e0"} err="failed to get container status \"39212d0b4bdff74a3a80cc7b73029334c760ccdbe73ce7e80ea61acba086b8e0\": rpc error: code = NotFound desc = could not find container \"39212d0b4bdff74a3a80cc7b73029334c760ccdbe73ce7e80ea61acba086b8e0\": container with ID starting with 39212d0b4bdff74a3a80cc7b73029334c760ccdbe73ce7e80ea61acba086b8e0 not found: ID does not exist" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.616180 4983 scope.go:117] "RemoveContainer" containerID="65f68d3e8aeb57fde85764cf459ca9da645245cc3bcb1dd9e27c8af9e88e0c3e" Mar 18 00:24:53 crc kubenswrapper[4983]: E0318 00:24:53.616743 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65f68d3e8aeb57fde85764cf459ca9da645245cc3bcb1dd9e27c8af9e88e0c3e\": container with ID starting with 65f68d3e8aeb57fde85764cf459ca9da645245cc3bcb1dd9e27c8af9e88e0c3e not found: ID does not exist" containerID="65f68d3e8aeb57fde85764cf459ca9da645245cc3bcb1dd9e27c8af9e88e0c3e" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.616810 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65f68d3e8aeb57fde85764cf459ca9da645245cc3bcb1dd9e27c8af9e88e0c3e"} err="failed to get container status \"65f68d3e8aeb57fde85764cf459ca9da645245cc3bcb1dd9e27c8af9e88e0c3e\": rpc error: code = NotFound desc = could not find container \"65f68d3e8aeb57fde85764cf459ca9da645245cc3bcb1dd9e27c8af9e88e0c3e\": container with ID starting with 65f68d3e8aeb57fde85764cf459ca9da645245cc3bcb1dd9e27c8af9e88e0c3e not found: ID does not exist" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.616850 4983 scope.go:117] "RemoveContainer" containerID="4f6520530aa53a878471e4e2687293b15e28ea02fc847a570e9de1f19c8f9252" Mar 18 00:24:53 crc kubenswrapper[4983]: E0318 00:24:53.617638 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f6520530aa53a878471e4e2687293b15e28ea02fc847a570e9de1f19c8f9252\": container with ID starting with 4f6520530aa53a878471e4e2687293b15e28ea02fc847a570e9de1f19c8f9252 not found: ID does not exist" containerID="4f6520530aa53a878471e4e2687293b15e28ea02fc847a570e9de1f19c8f9252" Mar 18 00:24:53 crc kubenswrapper[4983]: I0318 00:24:53.617709 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f6520530aa53a878471e4e2687293b15e28ea02fc847a570e9de1f19c8f9252"} err="failed to get container status \"4f6520530aa53a878471e4e2687293b15e28ea02fc847a570e9de1f19c8f9252\": rpc error: code = NotFound desc = could not find container \"4f6520530aa53a878471e4e2687293b15e28ea02fc847a570e9de1f19c8f9252\": container with ID starting with 4f6520530aa53a878471e4e2687293b15e28ea02fc847a570e9de1f19c8f9252 not found: ID does not exist" Mar 18 00:24:54 crc kubenswrapper[4983]: I0318 00:24:54.851888 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f01fa7bc-915f-4076-a5e8-8bc01bb745af" path="/var/lib/kubelet/pods/f01fa7bc-915f-4076-a5e8-8bc01bb745af/volumes" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.539177 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sg5wn"] Mar 18 00:25:08 crc kubenswrapper[4983]: E0318 00:25:08.541294 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f01fa7bc-915f-4076-a5e8-8bc01bb745af" containerName="extract-content" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.541354 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f01fa7bc-915f-4076-a5e8-8bc01bb745af" containerName="extract-content" Mar 18 00:25:08 crc kubenswrapper[4983]: E0318 00:25:08.541384 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f01fa7bc-915f-4076-a5e8-8bc01bb745af" containerName="extract-utilities" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.541399 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f01fa7bc-915f-4076-a5e8-8bc01bb745af" containerName="extract-utilities" Mar 18 00:25:08 crc kubenswrapper[4983]: E0318 00:25:08.541415 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e752de16-f363-433a-90a0-d9ac16b98fe4" containerName="docker-build" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.541427 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="e752de16-f363-433a-90a0-d9ac16b98fe4" containerName="docker-build" Mar 18 00:25:08 crc kubenswrapper[4983]: E0318 00:25:08.541446 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f01fa7bc-915f-4076-a5e8-8bc01bb745af" containerName="registry-server" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.541455 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="f01fa7bc-915f-4076-a5e8-8bc01bb745af" containerName="registry-server" Mar 18 00:25:08 crc kubenswrapper[4983]: E0318 00:25:08.541470 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e752de16-f363-433a-90a0-d9ac16b98fe4" containerName="manage-dockerfile" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.541482 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="e752de16-f363-433a-90a0-d9ac16b98fe4" containerName="manage-dockerfile" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.541770 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="f01fa7bc-915f-4076-a5e8-8bc01bb745af" containerName="registry-server" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.541795 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="e752de16-f363-433a-90a0-d9ac16b98fe4" containerName="docker-build" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.543071 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.557781 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sg5wn"] Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.718376 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-utilities\") pod \"community-operators-sg5wn\" (UID: \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\") " pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.718416 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-catalog-content\") pod \"community-operators-sg5wn\" (UID: \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\") " pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.718951 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx8cc\" (UniqueName: \"kubernetes.io/projected/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-kube-api-access-sx8cc\") pod \"community-operators-sg5wn\" (UID: \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\") " pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.820832 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-utilities\") pod \"community-operators-sg5wn\" (UID: \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\") " pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.821140 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-catalog-content\") pod \"community-operators-sg5wn\" (UID: \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\") " pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.821297 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx8cc\" (UniqueName: \"kubernetes.io/projected/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-kube-api-access-sx8cc\") pod \"community-operators-sg5wn\" (UID: \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\") " pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.821398 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-utilities\") pod \"community-operators-sg5wn\" (UID: \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\") " pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.821481 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-catalog-content\") pod \"community-operators-sg5wn\" (UID: \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\") " pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.849746 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx8cc\" (UniqueName: \"kubernetes.io/projected/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-kube-api-access-sx8cc\") pod \"community-operators-sg5wn\" (UID: \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\") " pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:08 crc kubenswrapper[4983]: I0318 00:25:08.861601 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:09 crc kubenswrapper[4983]: I0318 00:25:09.085638 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sg5wn"] Mar 18 00:25:09 crc kubenswrapper[4983]: W0318 00:25:09.090129 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5e1bcc7_48df_4318_88e8_4cf584e9ad99.slice/crio-1f576c3fffa6909ffc8f3e3f40a8efceae2e71db73445971dada9dc32d1b26d5 WatchSource:0}: Error finding container 1f576c3fffa6909ffc8f3e3f40a8efceae2e71db73445971dada9dc32d1b26d5: Status 404 returned error can't find the container with id 1f576c3fffa6909ffc8f3e3f40a8efceae2e71db73445971dada9dc32d1b26d5 Mar 18 00:25:09 crc kubenswrapper[4983]: I0318 00:25:09.652311 4983 generic.go:334] "Generic (PLEG): container finished" podID="a5e1bcc7-48df-4318-88e8-4cf584e9ad99" containerID="d294b8470326ca6b15fdc094215cf383cadc964e650a9ae359c86b36b055aec3" exitCode=0 Mar 18 00:25:09 crc kubenswrapper[4983]: I0318 00:25:09.652401 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5wn" event={"ID":"a5e1bcc7-48df-4318-88e8-4cf584e9ad99","Type":"ContainerDied","Data":"d294b8470326ca6b15fdc094215cf383cadc964e650a9ae359c86b36b055aec3"} Mar 18 00:25:09 crc kubenswrapper[4983]: I0318 00:25:09.652848 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5wn" event={"ID":"a5e1bcc7-48df-4318-88e8-4cf584e9ad99","Type":"ContainerStarted","Data":"1f576c3fffa6909ffc8f3e3f40a8efceae2e71db73445971dada9dc32d1b26d5"} Mar 18 00:25:10 crc kubenswrapper[4983]: I0318 00:25:10.663528 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5wn" event={"ID":"a5e1bcc7-48df-4318-88e8-4cf584e9ad99","Type":"ContainerStarted","Data":"7628c3ae9b25660519fb65fe1f14b8cdee1c64772df0e7f456616631fb7ab6b9"} Mar 18 00:25:11 crc kubenswrapper[4983]: I0318 00:25:11.674360 4983 generic.go:334] "Generic (PLEG): container finished" podID="a5e1bcc7-48df-4318-88e8-4cf584e9ad99" containerID="7628c3ae9b25660519fb65fe1f14b8cdee1c64772df0e7f456616631fb7ab6b9" exitCode=0 Mar 18 00:25:11 crc kubenswrapper[4983]: I0318 00:25:11.674402 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5wn" event={"ID":"a5e1bcc7-48df-4318-88e8-4cf584e9ad99","Type":"ContainerDied","Data":"7628c3ae9b25660519fb65fe1f14b8cdee1c64772df0e7f456616631fb7ab6b9"} Mar 18 00:25:12 crc kubenswrapper[4983]: I0318 00:25:12.704251 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5wn" event={"ID":"a5e1bcc7-48df-4318-88e8-4cf584e9ad99","Type":"ContainerStarted","Data":"f35859268d0f5a8009f03ddbc31b5634287a395b8d1c54d796fa0dbe29043670"} Mar 18 00:25:12 crc kubenswrapper[4983]: I0318 00:25:12.732559 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sg5wn" podStartSLOduration=2.326221569 podStartE2EDuration="4.732538628s" podCreationTimestamp="2026-03-18 00:25:08 +0000 UTC" firstStartedPulling="2026-03-18 00:25:09.654433342 +0000 UTC m=+1027.551807962" lastFinishedPulling="2026-03-18 00:25:12.060750351 +0000 UTC m=+1029.958125021" observedRunningTime="2026-03-18 00:25:12.727770703 +0000 UTC m=+1030.625145323" watchObservedRunningTime="2026-03-18 00:25:12.732538628 +0000 UTC m=+1030.629913248" Mar 18 00:25:18 crc kubenswrapper[4983]: I0318 00:25:18.863021 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:18 crc kubenswrapper[4983]: I0318 00:25:18.864751 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:18 crc kubenswrapper[4983]: I0318 00:25:18.930710 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:19 crc kubenswrapper[4983]: I0318 00:25:19.821034 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:19 crc kubenswrapper[4983]: I0318 00:25:19.882678 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sg5wn"] Mar 18 00:25:21 crc kubenswrapper[4983]: I0318 00:25:21.773488 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sg5wn" podUID="a5e1bcc7-48df-4318-88e8-4cf584e9ad99" containerName="registry-server" containerID="cri-o://f35859268d0f5a8009f03ddbc31b5634287a395b8d1c54d796fa0dbe29043670" gracePeriod=2 Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.201320 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.343433 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-utilities\") pod \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\" (UID: \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\") " Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.343535 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sx8cc\" (UniqueName: \"kubernetes.io/projected/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-kube-api-access-sx8cc\") pod \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\" (UID: \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\") " Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.343664 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-catalog-content\") pod \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\" (UID: \"a5e1bcc7-48df-4318-88e8-4cf584e9ad99\") " Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.345072 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-utilities" (OuterVolumeSpecName: "utilities") pod "a5e1bcc7-48df-4318-88e8-4cf584e9ad99" (UID: "a5e1bcc7-48df-4318-88e8-4cf584e9ad99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.351183 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-kube-api-access-sx8cc" (OuterVolumeSpecName: "kube-api-access-sx8cc") pod "a5e1bcc7-48df-4318-88e8-4cf584e9ad99" (UID: "a5e1bcc7-48df-4318-88e8-4cf584e9ad99"). InnerVolumeSpecName "kube-api-access-sx8cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.410341 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5e1bcc7-48df-4318-88e8-4cf584e9ad99" (UID: "a5e1bcc7-48df-4318-88e8-4cf584e9ad99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.446213 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.446706 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx8cc\" (UniqueName: \"kubernetes.io/projected/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-kube-api-access-sx8cc\") on node \"crc\" DevicePath \"\"" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.447043 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e1bcc7-48df-4318-88e8-4cf584e9ad99-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.783986 4983 generic.go:334] "Generic (PLEG): container finished" podID="a5e1bcc7-48df-4318-88e8-4cf584e9ad99" containerID="f35859268d0f5a8009f03ddbc31b5634287a395b8d1c54d796fa0dbe29043670" exitCode=0 Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.784027 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5wn" event={"ID":"a5e1bcc7-48df-4318-88e8-4cf584e9ad99","Type":"ContainerDied","Data":"f35859268d0f5a8009f03ddbc31b5634287a395b8d1c54d796fa0dbe29043670"} Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.784061 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5wn" event={"ID":"a5e1bcc7-48df-4318-88e8-4cf584e9ad99","Type":"ContainerDied","Data":"1f576c3fffa6909ffc8f3e3f40a8efceae2e71db73445971dada9dc32d1b26d5"} Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.784079 4983 scope.go:117] "RemoveContainer" containerID="f35859268d0f5a8009f03ddbc31b5634287a395b8d1c54d796fa0dbe29043670" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.785678 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sg5wn" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.806137 4983 scope.go:117] "RemoveContainer" containerID="7628c3ae9b25660519fb65fe1f14b8cdee1c64772df0e7f456616631fb7ab6b9" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.833170 4983 scope.go:117] "RemoveContainer" containerID="d294b8470326ca6b15fdc094215cf383cadc964e650a9ae359c86b36b055aec3" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.879383 4983 scope.go:117] "RemoveContainer" containerID="f35859268d0f5a8009f03ddbc31b5634287a395b8d1c54d796fa0dbe29043670" Mar 18 00:25:22 crc kubenswrapper[4983]: E0318 00:25:22.880672 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f35859268d0f5a8009f03ddbc31b5634287a395b8d1c54d796fa0dbe29043670\": container with ID starting with f35859268d0f5a8009f03ddbc31b5634287a395b8d1c54d796fa0dbe29043670 not found: ID does not exist" containerID="f35859268d0f5a8009f03ddbc31b5634287a395b8d1c54d796fa0dbe29043670" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.880739 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f35859268d0f5a8009f03ddbc31b5634287a395b8d1c54d796fa0dbe29043670"} err="failed to get container status \"f35859268d0f5a8009f03ddbc31b5634287a395b8d1c54d796fa0dbe29043670\": rpc error: code = NotFound desc = could not find container \"f35859268d0f5a8009f03ddbc31b5634287a395b8d1c54d796fa0dbe29043670\": container with ID starting with f35859268d0f5a8009f03ddbc31b5634287a395b8d1c54d796fa0dbe29043670 not found: ID does not exist" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.880787 4983 scope.go:117] "RemoveContainer" containerID="7628c3ae9b25660519fb65fe1f14b8cdee1c64772df0e7f456616631fb7ab6b9" Mar 18 00:25:22 crc kubenswrapper[4983]: E0318 00:25:22.881484 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7628c3ae9b25660519fb65fe1f14b8cdee1c64772df0e7f456616631fb7ab6b9\": container with ID starting with 7628c3ae9b25660519fb65fe1f14b8cdee1c64772df0e7f456616631fb7ab6b9 not found: ID does not exist" containerID="7628c3ae9b25660519fb65fe1f14b8cdee1c64772df0e7f456616631fb7ab6b9" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.881536 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7628c3ae9b25660519fb65fe1f14b8cdee1c64772df0e7f456616631fb7ab6b9"} err="failed to get container status \"7628c3ae9b25660519fb65fe1f14b8cdee1c64772df0e7f456616631fb7ab6b9\": rpc error: code = NotFound desc = could not find container \"7628c3ae9b25660519fb65fe1f14b8cdee1c64772df0e7f456616631fb7ab6b9\": container with ID starting with 7628c3ae9b25660519fb65fe1f14b8cdee1c64772df0e7f456616631fb7ab6b9 not found: ID does not exist" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.881572 4983 scope.go:117] "RemoveContainer" containerID="d294b8470326ca6b15fdc094215cf383cadc964e650a9ae359c86b36b055aec3" Mar 18 00:25:22 crc kubenswrapper[4983]: E0318 00:25:22.882270 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d294b8470326ca6b15fdc094215cf383cadc964e650a9ae359c86b36b055aec3\": container with ID starting with d294b8470326ca6b15fdc094215cf383cadc964e650a9ae359c86b36b055aec3 not found: ID does not exist" containerID="d294b8470326ca6b15fdc094215cf383cadc964e650a9ae359c86b36b055aec3" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.882440 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d294b8470326ca6b15fdc094215cf383cadc964e650a9ae359c86b36b055aec3"} err="failed to get container status \"d294b8470326ca6b15fdc094215cf383cadc964e650a9ae359c86b36b055aec3\": rpc error: code = NotFound desc = could not find container \"d294b8470326ca6b15fdc094215cf383cadc964e650a9ae359c86b36b055aec3\": container with ID starting with d294b8470326ca6b15fdc094215cf383cadc964e650a9ae359c86b36b055aec3 not found: ID does not exist" Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.884984 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sg5wn"] Mar 18 00:25:22 crc kubenswrapper[4983]: I0318 00:25:22.897917 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sg5wn"] Mar 18 00:25:24 crc kubenswrapper[4983]: I0318 00:25:24.857267 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5e1bcc7-48df-4318-88e8-4cf584e9ad99" path="/var/lib/kubelet/pods/a5e1bcc7-48df-4318-88e8-4cf584e9ad99/volumes" Mar 18 00:25:59 crc kubenswrapper[4983]: I0318 00:25:59.058127 4983 generic.go:334] "Generic (PLEG): container finished" podID="a507119f-75ce-48dc-ba8d-374e9c39cd09" containerID="5636a5b2375cfdbaae981b745551fae086f4c3a9b4777b501f5f8a4649de5d3e" exitCode=0 Mar 18 00:25:59 crc kubenswrapper[4983]: I0318 00:25:59.058207 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a507119f-75ce-48dc-ba8d-374e9c39cd09","Type":"ContainerDied","Data":"5636a5b2375cfdbaae981b745551fae086f4c3a9b4777b501f5f8a4649de5d3e"} Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.143956 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563226-87lzm"] Mar 18 00:26:00 crc kubenswrapper[4983]: E0318 00:26:00.144180 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e1bcc7-48df-4318-88e8-4cf584e9ad99" containerName="registry-server" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.144191 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e1bcc7-48df-4318-88e8-4cf584e9ad99" containerName="registry-server" Mar 18 00:26:00 crc kubenswrapper[4983]: E0318 00:26:00.144415 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e1bcc7-48df-4318-88e8-4cf584e9ad99" containerName="extract-utilities" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.144421 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e1bcc7-48df-4318-88e8-4cf584e9ad99" containerName="extract-utilities" Mar 18 00:26:00 crc kubenswrapper[4983]: E0318 00:26:00.144435 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e1bcc7-48df-4318-88e8-4cf584e9ad99" containerName="extract-content" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.144441 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e1bcc7-48df-4318-88e8-4cf584e9ad99" containerName="extract-content" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.144545 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5e1bcc7-48df-4318-88e8-4cf584e9ad99" containerName="registry-server" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.144947 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563226-87lzm" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.147464 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.147520 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.147614 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.203034 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563226-87lzm"] Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.341254 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px5hs\" (UniqueName: \"kubernetes.io/projected/018dc6e9-49c7-443a-8f85-854247ef9d94-kube-api-access-px5hs\") pod \"auto-csr-approver-29563226-87lzm\" (UID: \"018dc6e9-49c7-443a-8f85-854247ef9d94\") " pod="openshift-infra/auto-csr-approver-29563226-87lzm" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.404961 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.442009 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-container-storage-root\") pod \"a507119f-75ce-48dc-ba8d-374e9c39cd09\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.442052 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/a507119f-75ce-48dc-ba8d-374e9c39cd09-builder-dockercfg-q2gzr-push\") pod \"a507119f-75ce-48dc-ba8d-374e9c39cd09\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.442107 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/a507119f-75ce-48dc-ba8d-374e9c39cd09-builder-dockercfg-q2gzr-pull\") pod \"a507119f-75ce-48dc-ba8d-374e9c39cd09\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.442133 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-ca-bundles\") pod \"a507119f-75ce-48dc-ba8d-374e9c39cd09\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.442151 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-blob-cache\") pod \"a507119f-75ce-48dc-ba8d-374e9c39cd09\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.442191 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-buildworkdir\") pod \"a507119f-75ce-48dc-ba8d-374e9c39cd09\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.442211 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a507119f-75ce-48dc-ba8d-374e9c39cd09-node-pullsecrets\") pod \"a507119f-75ce-48dc-ba8d-374e9c39cd09\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.442234 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-proxy-ca-bundles\") pod \"a507119f-75ce-48dc-ba8d-374e9c39cd09\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.442262 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqgr2\" (UniqueName: \"kubernetes.io/projected/a507119f-75ce-48dc-ba8d-374e9c39cd09-kube-api-access-wqgr2\") pod \"a507119f-75ce-48dc-ba8d-374e9c39cd09\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.442278 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-system-configs\") pod \"a507119f-75ce-48dc-ba8d-374e9c39cd09\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.442293 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-container-storage-run\") pod \"a507119f-75ce-48dc-ba8d-374e9c39cd09\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.442309 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a507119f-75ce-48dc-ba8d-374e9c39cd09-buildcachedir\") pod \"a507119f-75ce-48dc-ba8d-374e9c39cd09\" (UID: \"a507119f-75ce-48dc-ba8d-374e9c39cd09\") " Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.442392 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px5hs\" (UniqueName: \"kubernetes.io/projected/018dc6e9-49c7-443a-8f85-854247ef9d94-kube-api-access-px5hs\") pod \"auto-csr-approver-29563226-87lzm\" (UID: \"018dc6e9-49c7-443a-8f85-854247ef9d94\") " pod="openshift-infra/auto-csr-approver-29563226-87lzm" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.443727 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a507119f-75ce-48dc-ba8d-374e9c39cd09-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "a507119f-75ce-48dc-ba8d-374e9c39cd09" (UID: "a507119f-75ce-48dc-ba8d-374e9c39cd09"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.443868 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a507119f-75ce-48dc-ba8d-374e9c39cd09-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "a507119f-75ce-48dc-ba8d-374e9c39cd09" (UID: "a507119f-75ce-48dc-ba8d-374e9c39cd09"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.444572 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "a507119f-75ce-48dc-ba8d-374e9c39cd09" (UID: "a507119f-75ce-48dc-ba8d-374e9c39cd09"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.445108 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "a507119f-75ce-48dc-ba8d-374e9c39cd09" (UID: "a507119f-75ce-48dc-ba8d-374e9c39cd09"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.445170 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "a507119f-75ce-48dc-ba8d-374e9c39cd09" (UID: "a507119f-75ce-48dc-ba8d-374e9c39cd09"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.445227 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "a507119f-75ce-48dc-ba8d-374e9c39cd09" (UID: "a507119f-75ce-48dc-ba8d-374e9c39cd09"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.451823 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a507119f-75ce-48dc-ba8d-374e9c39cd09-builder-dockercfg-q2gzr-pull" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-pull") pod "a507119f-75ce-48dc-ba8d-374e9c39cd09" (UID: "a507119f-75ce-48dc-ba8d-374e9c39cd09"). InnerVolumeSpecName "builder-dockercfg-q2gzr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.459901 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a507119f-75ce-48dc-ba8d-374e9c39cd09-builder-dockercfg-q2gzr-push" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-push") pod "a507119f-75ce-48dc-ba8d-374e9c39cd09" (UID: "a507119f-75ce-48dc-ba8d-374e9c39cd09"). InnerVolumeSpecName "builder-dockercfg-q2gzr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.461070 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a507119f-75ce-48dc-ba8d-374e9c39cd09-kube-api-access-wqgr2" (OuterVolumeSpecName: "kube-api-access-wqgr2") pod "a507119f-75ce-48dc-ba8d-374e9c39cd09" (UID: "a507119f-75ce-48dc-ba8d-374e9c39cd09"). InnerVolumeSpecName "kube-api-access-wqgr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.462604 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "a507119f-75ce-48dc-ba8d-374e9c39cd09" (UID: "a507119f-75ce-48dc-ba8d-374e9c39cd09"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.469461 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px5hs\" (UniqueName: \"kubernetes.io/projected/018dc6e9-49c7-443a-8f85-854247ef9d94-kube-api-access-px5hs\") pod \"auto-csr-approver-29563226-87lzm\" (UID: \"018dc6e9-49c7-443a-8f85-854247ef9d94\") " pod="openshift-infra/auto-csr-approver-29563226-87lzm" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.545844 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqgr2\" (UniqueName: \"kubernetes.io/projected/a507119f-75ce-48dc-ba8d-374e9c39cd09-kube-api-access-wqgr2\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.545900 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.545921 4983 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.545973 4983 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a507119f-75ce-48dc-ba8d-374e9c39cd09-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.545995 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/a507119f-75ce-48dc-ba8d-374e9c39cd09-builder-dockercfg-q2gzr-push\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.546049 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/a507119f-75ce-48dc-ba8d-374e9c39cd09-builder-dockercfg-q2gzr-pull\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.546068 4983 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.546092 4983 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.546114 4983 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a507119f-75ce-48dc-ba8d-374e9c39cd09-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.546136 4983 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.758876 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563226-87lzm" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.902946 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "a507119f-75ce-48dc-ba8d-374e9c39cd09" (UID: "a507119f-75ce-48dc-ba8d-374e9c39cd09"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:26:00 crc kubenswrapper[4983]: I0318 00:26:00.959984 4983 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:01 crc kubenswrapper[4983]: I0318 00:26:01.003292 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563226-87lzm"] Mar 18 00:26:01 crc kubenswrapper[4983]: I0318 00:26:01.074039 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a507119f-75ce-48dc-ba8d-374e9c39cd09","Type":"ContainerDied","Data":"c2a3b3a058c50f42dc7b9c24a1590b42694128f5fa1053a6b122de4c19db241e"} Mar 18 00:26:01 crc kubenswrapper[4983]: I0318 00:26:01.074103 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2a3b3a058c50f42dc7b9c24a1590b42694128f5fa1053a6b122de4c19db241e" Mar 18 00:26:01 crc kubenswrapper[4983]: I0318 00:26:01.074292 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Mar 18 00:26:01 crc kubenswrapper[4983]: I0318 00:26:01.076408 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563226-87lzm" event={"ID":"018dc6e9-49c7-443a-8f85-854247ef9d94","Type":"ContainerStarted","Data":"8f7fc42fc2049edd9729223516c337d367fa8ea877d5fb1ab2cce11e0ab97728"} Mar 18 00:26:02 crc kubenswrapper[4983]: I0318 00:26:02.818065 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "a507119f-75ce-48dc-ba8d-374e9c39cd09" (UID: "a507119f-75ce-48dc-ba8d-374e9c39cd09"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:26:02 crc kubenswrapper[4983]: I0318 00:26:02.891153 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a507119f-75ce-48dc-ba8d-374e9c39cd09-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:03 crc kubenswrapper[4983]: I0318 00:26:03.091305 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563226-87lzm" event={"ID":"018dc6e9-49c7-443a-8f85-854247ef9d94","Type":"ContainerStarted","Data":"5553c9f8f36da1a7fa7e9e2d4fe6285578f7857a0609e4f1eb6ba3eccc6c0bc4"} Mar 18 00:26:03 crc kubenswrapper[4983]: I0318 00:26:03.111895 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29563226-87lzm" podStartSLOduration=1.470511578 podStartE2EDuration="3.111878218s" podCreationTimestamp="2026-03-18 00:26:00 +0000 UTC" firstStartedPulling="2026-03-18 00:26:01.004630492 +0000 UTC m=+1078.902005102" lastFinishedPulling="2026-03-18 00:26:02.645997122 +0000 UTC m=+1080.543371742" observedRunningTime="2026-03-18 00:26:03.107109603 +0000 UTC m=+1081.004484223" watchObservedRunningTime="2026-03-18 00:26:03.111878218 +0000 UTC m=+1081.009252838" Mar 18 00:26:04 crc kubenswrapper[4983]: I0318 00:26:04.099969 4983 generic.go:334] "Generic (PLEG): container finished" podID="018dc6e9-49c7-443a-8f85-854247ef9d94" containerID="5553c9f8f36da1a7fa7e9e2d4fe6285578f7857a0609e4f1eb6ba3eccc6c0bc4" exitCode=0 Mar 18 00:26:04 crc kubenswrapper[4983]: I0318 00:26:04.100014 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563226-87lzm" event={"ID":"018dc6e9-49c7-443a-8f85-854247ef9d94","Type":"ContainerDied","Data":"5553c9f8f36da1a7fa7e9e2d4fe6285578f7857a0609e4f1eb6ba3eccc6c0bc4"} Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.308265 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 18 00:26:05 crc kubenswrapper[4983]: E0318 00:26:05.308762 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a507119f-75ce-48dc-ba8d-374e9c39cd09" containerName="docker-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.308776 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a507119f-75ce-48dc-ba8d-374e9c39cd09" containerName="docker-build" Mar 18 00:26:05 crc kubenswrapper[4983]: E0318 00:26:05.308794 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a507119f-75ce-48dc-ba8d-374e9c39cd09" containerName="manage-dockerfile" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.308801 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a507119f-75ce-48dc-ba8d-374e9c39cd09" containerName="manage-dockerfile" Mar 18 00:26:05 crc kubenswrapper[4983]: E0318 00:26:05.308814 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a507119f-75ce-48dc-ba8d-374e9c39cd09" containerName="git-clone" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.308824 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="a507119f-75ce-48dc-ba8d-374e9c39cd09" containerName="git-clone" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.308974 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="a507119f-75ce-48dc-ba8d-374e9c39cd09" containerName="docker-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.309677 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.312421 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-global-ca" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.312713 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-ca" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.312893 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-sys-config" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.313031 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-q2gzr" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.323036 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.334990 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b351526d-7b1d-4fe3-94f8-2a274254c42c-buildcachedir\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.335034 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-buildworkdir\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.362704 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563226-87lzm" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.435960 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.436069 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-container-storage-root\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.436165 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.436226 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-container-storage-run\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.436313 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.436366 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b351526d-7b1d-4fe3-94f8-2a274254c42c-buildcachedir\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.436399 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/b351526d-7b1d-4fe3-94f8-2a274254c42c-builder-dockercfg-q2gzr-push\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.436899 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnbqf\" (UniqueName: \"kubernetes.io/projected/b351526d-7b1d-4fe3-94f8-2a274254c42c-kube-api-access-lnbqf\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.436937 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b351526d-7b1d-4fe3-94f8-2a274254c42c-buildcachedir\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.436953 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-buildworkdir\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.437366 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-buildworkdir\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.437441 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b351526d-7b1d-4fe3-94f8-2a274254c42c-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.437528 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/b351526d-7b1d-4fe3-94f8-2a274254c42c-builder-dockercfg-q2gzr-pull\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.437557 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-system-configs\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.538307 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px5hs\" (UniqueName: \"kubernetes.io/projected/018dc6e9-49c7-443a-8f85-854247ef9d94-kube-api-access-px5hs\") pod \"018dc6e9-49c7-443a-8f85-854247ef9d94\" (UID: \"018dc6e9-49c7-443a-8f85-854247ef9d94\") " Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.538572 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/b351526d-7b1d-4fe3-94f8-2a274254c42c-builder-dockercfg-q2gzr-push\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.538648 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnbqf\" (UniqueName: \"kubernetes.io/projected/b351526d-7b1d-4fe3-94f8-2a274254c42c-kube-api-access-lnbqf\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.538703 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b351526d-7b1d-4fe3-94f8-2a274254c42c-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.538782 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/b351526d-7b1d-4fe3-94f8-2a274254c42c-builder-dockercfg-q2gzr-pull\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.538998 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b351526d-7b1d-4fe3-94f8-2a274254c42c-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.539260 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-system-configs\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.540024 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.540111 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-container-storage-root\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.540180 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.540265 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-system-configs\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.540803 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-container-storage-root\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.540905 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-container-storage-run\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.540918 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.541066 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.541293 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-container-storage-run\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.542073 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.542374 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.548269 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/018dc6e9-49c7-443a-8f85-854247ef9d94-kube-api-access-px5hs" (OuterVolumeSpecName: "kube-api-access-px5hs") pod "018dc6e9-49c7-443a-8f85-854247ef9d94" (UID: "018dc6e9-49c7-443a-8f85-854247ef9d94"). InnerVolumeSpecName "kube-api-access-px5hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.548855 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/b351526d-7b1d-4fe3-94f8-2a274254c42c-builder-dockercfg-q2gzr-push\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.551984 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/b351526d-7b1d-4fe3-94f8-2a274254c42c-builder-dockercfg-q2gzr-pull\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.568060 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnbqf\" (UniqueName: \"kubernetes.io/projected/b351526d-7b1d-4fe3-94f8-2a274254c42c-kube-api-access-lnbqf\") pod \"sg-core-1-build\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.642700 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px5hs\" (UniqueName: \"kubernetes.io/projected/018dc6e9-49c7-443a-8f85-854247ef9d94-kube-api-access-px5hs\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.660780 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.934993 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563220-zdtdg"] Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.946325 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563220-zdtdg"] Mar 18 00:26:05 crc kubenswrapper[4983]: I0318 00:26:05.959128 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 18 00:26:06 crc kubenswrapper[4983]: I0318 00:26:06.115723 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563226-87lzm" event={"ID":"018dc6e9-49c7-443a-8f85-854247ef9d94","Type":"ContainerDied","Data":"8f7fc42fc2049edd9729223516c337d367fa8ea877d5fb1ab2cce11e0ab97728"} Mar 18 00:26:06 crc kubenswrapper[4983]: I0318 00:26:06.115769 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f7fc42fc2049edd9729223516c337d367fa8ea877d5fb1ab2cce11e0ab97728" Mar 18 00:26:06 crc kubenswrapper[4983]: I0318 00:26:06.115783 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563226-87lzm" Mar 18 00:26:06 crc kubenswrapper[4983]: I0318 00:26:06.116788 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"b351526d-7b1d-4fe3-94f8-2a274254c42c","Type":"ContainerStarted","Data":"9ec2c5b80e9674611e27684fbef104c6951bad7dd3abb47e174372e06c4d2b32"} Mar 18 00:26:06 crc kubenswrapper[4983]: I0318 00:26:06.869680 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d149429e-e5d1-48e5-a052-feecd610f7c5" path="/var/lib/kubelet/pods/d149429e-e5d1-48e5-a052-feecd610f7c5/volumes" Mar 18 00:26:07 crc kubenswrapper[4983]: I0318 00:26:07.135842 4983 generic.go:334] "Generic (PLEG): container finished" podID="b351526d-7b1d-4fe3-94f8-2a274254c42c" containerID="53ca1479e1cc8045375be084f8f4aa96b6dc4d2d328f286b9f352fb2e7070cc2" exitCode=0 Mar 18 00:26:07 crc kubenswrapper[4983]: I0318 00:26:07.135904 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"b351526d-7b1d-4fe3-94f8-2a274254c42c","Type":"ContainerDied","Data":"53ca1479e1cc8045375be084f8f4aa96b6dc4d2d328f286b9f352fb2e7070cc2"} Mar 18 00:26:08 crc kubenswrapper[4983]: I0318 00:26:08.148548 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"b351526d-7b1d-4fe3-94f8-2a274254c42c","Type":"ContainerStarted","Data":"93d22406581f7a9ed3f47754d929fd1684e54e6fdf4768c79595fc9afeb9ae65"} Mar 18 00:26:08 crc kubenswrapper[4983]: I0318 00:26:08.197649 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-1-build" podStartSLOduration=3.197626354 podStartE2EDuration="3.197626354s" podCreationTimestamp="2026-03-18 00:26:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:26:08.189396501 +0000 UTC m=+1086.086771171" watchObservedRunningTime="2026-03-18 00:26:08.197626354 +0000 UTC m=+1086.095000984" Mar 18 00:26:09 crc kubenswrapper[4983]: I0318 00:26:09.926323 4983 scope.go:117] "RemoveContainer" containerID="41437bb0365646c4a12e203f89ea7a4caa43e57dda1e59a4bb40f5a9e4bae3fe" Mar 18 00:26:15 crc kubenswrapper[4983]: I0318 00:26:15.683380 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 18 00:26:15 crc kubenswrapper[4983]: I0318 00:26:15.684559 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/sg-core-1-build" podUID="b351526d-7b1d-4fe3-94f8-2a274254c42c" containerName="docker-build" containerID="cri-o://93d22406581f7a9ed3f47754d929fd1684e54e6fdf4768c79595fc9afeb9ae65" gracePeriod=30 Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.155468 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_b351526d-7b1d-4fe3-94f8-2a274254c42c/docker-build/0.log" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.156386 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.196713 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-blob-cache\") pod \"b351526d-7b1d-4fe3-94f8-2a274254c42c\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.196804 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-system-configs\") pod \"b351526d-7b1d-4fe3-94f8-2a274254c42c\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.196849 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-container-storage-root\") pod \"b351526d-7b1d-4fe3-94f8-2a274254c42c\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.196909 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b351526d-7b1d-4fe3-94f8-2a274254c42c-node-pullsecrets\") pod \"b351526d-7b1d-4fe3-94f8-2a274254c42c\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.197050 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/b351526d-7b1d-4fe3-94f8-2a274254c42c-builder-dockercfg-q2gzr-pull\") pod \"b351526d-7b1d-4fe3-94f8-2a274254c42c\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.197091 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnbqf\" (UniqueName: \"kubernetes.io/projected/b351526d-7b1d-4fe3-94f8-2a274254c42c-kube-api-access-lnbqf\") pod \"b351526d-7b1d-4fe3-94f8-2a274254c42c\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.197149 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/b351526d-7b1d-4fe3-94f8-2a274254c42c-builder-dockercfg-q2gzr-push\") pod \"b351526d-7b1d-4fe3-94f8-2a274254c42c\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.197195 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-buildworkdir\") pod \"b351526d-7b1d-4fe3-94f8-2a274254c42c\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.197228 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-ca-bundles\") pod \"b351526d-7b1d-4fe3-94f8-2a274254c42c\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.197261 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-proxy-ca-bundles\") pod \"b351526d-7b1d-4fe3-94f8-2a274254c42c\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.197309 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b351526d-7b1d-4fe3-94f8-2a274254c42c-buildcachedir\") pod \"b351526d-7b1d-4fe3-94f8-2a274254c42c\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.197386 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-container-storage-run\") pod \"b351526d-7b1d-4fe3-94f8-2a274254c42c\" (UID: \"b351526d-7b1d-4fe3-94f8-2a274254c42c\") " Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.197636 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b351526d-7b1d-4fe3-94f8-2a274254c42c-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "b351526d-7b1d-4fe3-94f8-2a274254c42c" (UID: "b351526d-7b1d-4fe3-94f8-2a274254c42c"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.197867 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b351526d-7b1d-4fe3-94f8-2a274254c42c-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "b351526d-7b1d-4fe3-94f8-2a274254c42c" (UID: "b351526d-7b1d-4fe3-94f8-2a274254c42c"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.197950 4983 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b351526d-7b1d-4fe3-94f8-2a274254c42c-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.198653 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "b351526d-7b1d-4fe3-94f8-2a274254c42c" (UID: "b351526d-7b1d-4fe3-94f8-2a274254c42c"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.201040 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "b351526d-7b1d-4fe3-94f8-2a274254c42c" (UID: "b351526d-7b1d-4fe3-94f8-2a274254c42c"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.205703 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "b351526d-7b1d-4fe3-94f8-2a274254c42c" (UID: "b351526d-7b1d-4fe3-94f8-2a274254c42c"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.207658 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "b351526d-7b1d-4fe3-94f8-2a274254c42c" (UID: "b351526d-7b1d-4fe3-94f8-2a274254c42c"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.211195 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b351526d-7b1d-4fe3-94f8-2a274254c42c-builder-dockercfg-q2gzr-push" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-push") pod "b351526d-7b1d-4fe3-94f8-2a274254c42c" (UID: "b351526d-7b1d-4fe3-94f8-2a274254c42c"). InnerVolumeSpecName "builder-dockercfg-q2gzr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.218093 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "b351526d-7b1d-4fe3-94f8-2a274254c42c" (UID: "b351526d-7b1d-4fe3-94f8-2a274254c42c"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.224055 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b351526d-7b1d-4fe3-94f8-2a274254c42c-builder-dockercfg-q2gzr-pull" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-pull") pod "b351526d-7b1d-4fe3-94f8-2a274254c42c" (UID: "b351526d-7b1d-4fe3-94f8-2a274254c42c"). InnerVolumeSpecName "builder-dockercfg-q2gzr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.226043 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b351526d-7b1d-4fe3-94f8-2a274254c42c-kube-api-access-lnbqf" (OuterVolumeSpecName: "kube-api-access-lnbqf") pod "b351526d-7b1d-4fe3-94f8-2a274254c42c" (UID: "b351526d-7b1d-4fe3-94f8-2a274254c42c"). InnerVolumeSpecName "kube-api-access-lnbqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.259206 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_b351526d-7b1d-4fe3-94f8-2a274254c42c/docker-build/0.log" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.260554 4983 generic.go:334] "Generic (PLEG): container finished" podID="b351526d-7b1d-4fe3-94f8-2a274254c42c" containerID="93d22406581f7a9ed3f47754d929fd1684e54e6fdf4768c79595fc9afeb9ae65" exitCode=1 Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.260676 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"b351526d-7b1d-4fe3-94f8-2a274254c42c","Type":"ContainerDied","Data":"93d22406581f7a9ed3f47754d929fd1684e54e6fdf4768c79595fc9afeb9ae65"} Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.260763 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"b351526d-7b1d-4fe3-94f8-2a274254c42c","Type":"ContainerDied","Data":"9ec2c5b80e9674611e27684fbef104c6951bad7dd3abb47e174372e06c4d2b32"} Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.260845 4983 scope.go:117] "RemoveContainer" containerID="93d22406581f7a9ed3f47754d929fd1684e54e6fdf4768c79595fc9afeb9ae65" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.261085 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.299412 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.299460 4983 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.299474 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/b351526d-7b1d-4fe3-94f8-2a274254c42c-builder-dockercfg-q2gzr-pull\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.299495 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnbqf\" (UniqueName: \"kubernetes.io/projected/b351526d-7b1d-4fe3-94f8-2a274254c42c-kube-api-access-lnbqf\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.299505 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/b351526d-7b1d-4fe3-94f8-2a274254c42c-builder-dockercfg-q2gzr-push\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.299516 4983 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.299526 4983 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.299536 4983 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.299545 4983 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b351526d-7b1d-4fe3-94f8-2a274254c42c-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.351746 4983 scope.go:117] "RemoveContainer" containerID="53ca1479e1cc8045375be084f8f4aa96b6dc4d2d328f286b9f352fb2e7070cc2" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.378964 4983 scope.go:117] "RemoveContainer" containerID="93d22406581f7a9ed3f47754d929fd1684e54e6fdf4768c79595fc9afeb9ae65" Mar 18 00:26:16 crc kubenswrapper[4983]: E0318 00:26:16.379212 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93d22406581f7a9ed3f47754d929fd1684e54e6fdf4768c79595fc9afeb9ae65\": container with ID starting with 93d22406581f7a9ed3f47754d929fd1684e54e6fdf4768c79595fc9afeb9ae65 not found: ID does not exist" containerID="93d22406581f7a9ed3f47754d929fd1684e54e6fdf4768c79595fc9afeb9ae65" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.379245 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d22406581f7a9ed3f47754d929fd1684e54e6fdf4768c79595fc9afeb9ae65"} err="failed to get container status \"93d22406581f7a9ed3f47754d929fd1684e54e6fdf4768c79595fc9afeb9ae65\": rpc error: code = NotFound desc = could not find container \"93d22406581f7a9ed3f47754d929fd1684e54e6fdf4768c79595fc9afeb9ae65\": container with ID starting with 93d22406581f7a9ed3f47754d929fd1684e54e6fdf4768c79595fc9afeb9ae65 not found: ID does not exist" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.379265 4983 scope.go:117] "RemoveContainer" containerID="53ca1479e1cc8045375be084f8f4aa96b6dc4d2d328f286b9f352fb2e7070cc2" Mar 18 00:26:16 crc kubenswrapper[4983]: E0318 00:26:16.379641 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53ca1479e1cc8045375be084f8f4aa96b6dc4d2d328f286b9f352fb2e7070cc2\": container with ID starting with 53ca1479e1cc8045375be084f8f4aa96b6dc4d2d328f286b9f352fb2e7070cc2 not found: ID does not exist" containerID="53ca1479e1cc8045375be084f8f4aa96b6dc4d2d328f286b9f352fb2e7070cc2" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.379660 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53ca1479e1cc8045375be084f8f4aa96b6dc4d2d328f286b9f352fb2e7070cc2"} err="failed to get container status \"53ca1479e1cc8045375be084f8f4aa96b6dc4d2d328f286b9f352fb2e7070cc2\": rpc error: code = NotFound desc = could not find container \"53ca1479e1cc8045375be084f8f4aa96b6dc4d2d328f286b9f352fb2e7070cc2\": container with ID starting with 53ca1479e1cc8045375be084f8f4aa96b6dc4d2d328f286b9f352fb2e7070cc2 not found: ID does not exist" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.404386 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "b351526d-7b1d-4fe3-94f8-2a274254c42c" (UID: "b351526d-7b1d-4fe3-94f8-2a274254c42c"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.407379 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "b351526d-7b1d-4fe3-94f8-2a274254c42c" (UID: "b351526d-7b1d-4fe3-94f8-2a274254c42c"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.501632 4983 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.501697 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b351526d-7b1d-4fe3-94f8-2a274254c42c-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.612141 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.627222 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 18 00:26:16 crc kubenswrapper[4983]: I0318 00:26:16.859057 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b351526d-7b1d-4fe3-94f8-2a274254c42c" path="/var/lib/kubelet/pods/b351526d-7b1d-4fe3-94f8-2a274254c42c/volumes" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.325831 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-2-build"] Mar 18 00:26:17 crc kubenswrapper[4983]: E0318 00:26:17.326645 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b351526d-7b1d-4fe3-94f8-2a274254c42c" containerName="manage-dockerfile" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.326689 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="b351526d-7b1d-4fe3-94f8-2a274254c42c" containerName="manage-dockerfile" Mar 18 00:26:17 crc kubenswrapper[4983]: E0318 00:26:17.326720 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018dc6e9-49c7-443a-8f85-854247ef9d94" containerName="oc" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.326737 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="018dc6e9-49c7-443a-8f85-854247ef9d94" containerName="oc" Mar 18 00:26:17 crc kubenswrapper[4983]: E0318 00:26:17.326762 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b351526d-7b1d-4fe3-94f8-2a274254c42c" containerName="docker-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.326779 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="b351526d-7b1d-4fe3-94f8-2a274254c42c" containerName="docker-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.327075 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="b351526d-7b1d-4fe3-94f8-2a274254c42c" containerName="docker-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.327128 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="018dc6e9-49c7-443a-8f85-854247ef9d94" containerName="oc" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.331248 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.347125 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-sys-config" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.347388 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-global-ca" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.347760 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-q2gzr" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.352239 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-ca" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.375612 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.516342 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.516410 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.516435 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/e3f8b858-5134-424f-aa3f-fce238faead3-builder-dockercfg-q2gzr-pull\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.516456 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e3f8b858-5134-424f-aa3f-fce238faead3-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.516477 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e3f8b858-5134-424f-aa3f-fce238faead3-buildcachedir\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.516492 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.516536 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/e3f8b858-5134-424f-aa3f-fce238faead3-builder-dockercfg-q2gzr-push\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.516558 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-buildworkdir\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.516578 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9575w\" (UniqueName: \"kubernetes.io/projected/e3f8b858-5134-424f-aa3f-fce238faead3-kube-api-access-9575w\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.516594 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-container-storage-root\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.516634 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-system-configs\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.516659 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-container-storage-run\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.618093 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-container-storage-run\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.618213 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.618312 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.618427 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/e3f8b858-5134-424f-aa3f-fce238faead3-builder-dockercfg-q2gzr-pull\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.618850 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.619132 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-container-storage-run\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.619607 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e3f8b858-5134-424f-aa3f-fce238faead3-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.619680 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.619736 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e3f8b858-5134-424f-aa3f-fce238faead3-buildcachedir\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.619743 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e3f8b858-5134-424f-aa3f-fce238faead3-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.619769 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.619899 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e3f8b858-5134-424f-aa3f-fce238faead3-buildcachedir\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.620002 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/e3f8b858-5134-424f-aa3f-fce238faead3-builder-dockercfg-q2gzr-push\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.620093 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-buildworkdir\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.620157 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9575w\" (UniqueName: \"kubernetes.io/projected/e3f8b858-5134-424f-aa3f-fce238faead3-kube-api-access-9575w\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.620217 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-container-storage-root\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.620350 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-system-configs\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.620496 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.620736 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-buildworkdir\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.620847 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-container-storage-root\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.621304 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-system-configs\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.624674 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/e3f8b858-5134-424f-aa3f-fce238faead3-builder-dockercfg-q2gzr-pull\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.624784 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/e3f8b858-5134-424f-aa3f-fce238faead3-builder-dockercfg-q2gzr-push\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.638001 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9575w\" (UniqueName: \"kubernetes.io/projected/e3f8b858-5134-424f-aa3f-fce238faead3-kube-api-access-9575w\") pod \"sg-core-2-build\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.681616 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Mar 18 00:26:17 crc kubenswrapper[4983]: I0318 00:26:17.986452 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Mar 18 00:26:18 crc kubenswrapper[4983]: I0318 00:26:18.280462 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"e3f8b858-5134-424f-aa3f-fce238faead3","Type":"ContainerStarted","Data":"d26f5e0f286c0e6fa86a1261a0993d3353b2e2bd618fac4ca363bdc9ab680269"} Mar 18 00:26:19 crc kubenswrapper[4983]: I0318 00:26:19.292268 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"e3f8b858-5134-424f-aa3f-fce238faead3","Type":"ContainerStarted","Data":"1268c7a6e909d3cbeac652225cd5e6a9bb9f4f295c0519e2d6adc7f333dd53a7"} Mar 18 00:26:20 crc kubenswrapper[4983]: I0318 00:26:20.302993 4983 generic.go:334] "Generic (PLEG): container finished" podID="e3f8b858-5134-424f-aa3f-fce238faead3" containerID="1268c7a6e909d3cbeac652225cd5e6a9bb9f4f295c0519e2d6adc7f333dd53a7" exitCode=0 Mar 18 00:26:20 crc kubenswrapper[4983]: I0318 00:26:20.303111 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"e3f8b858-5134-424f-aa3f-fce238faead3","Type":"ContainerDied","Data":"1268c7a6e909d3cbeac652225cd5e6a9bb9f4f295c0519e2d6adc7f333dd53a7"} Mar 18 00:26:21 crc kubenswrapper[4983]: I0318 00:26:21.318284 4983 generic.go:334] "Generic (PLEG): container finished" podID="e3f8b858-5134-424f-aa3f-fce238faead3" containerID="c0b8f98f6e4b6e0e031e4da440549659a5e80cf991c2302d9bc63e41fa8f1b90" exitCode=0 Mar 18 00:26:21 crc kubenswrapper[4983]: I0318 00:26:21.318353 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"e3f8b858-5134-424f-aa3f-fce238faead3","Type":"ContainerDied","Data":"c0b8f98f6e4b6e0e031e4da440549659a5e80cf991c2302d9bc63e41fa8f1b90"} Mar 18 00:26:21 crc kubenswrapper[4983]: I0318 00:26:21.371666 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-2-build_e3f8b858-5134-424f-aa3f-fce238faead3/manage-dockerfile/0.log" Mar 18 00:26:22 crc kubenswrapper[4983]: I0318 00:26:22.335737 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"e3f8b858-5134-424f-aa3f-fce238faead3","Type":"ContainerStarted","Data":"b8339df555734c7201081ab92dc2a7031ec373b4c37ad72fb7f1d9c0360f091e"} Mar 18 00:26:22 crc kubenswrapper[4983]: I0318 00:26:22.391040 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-2-build" podStartSLOduration=5.391013925 podStartE2EDuration="5.391013925s" podCreationTimestamp="2026-03-18 00:26:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:26:22.377466071 +0000 UTC m=+1100.274840771" watchObservedRunningTime="2026-03-18 00:26:22.391013925 +0000 UTC m=+1100.288388565" Mar 18 00:26:43 crc kubenswrapper[4983]: I0318 00:26:43.259490 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:26:43 crc kubenswrapper[4983]: I0318 00:26:43.260015 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:27:13 crc kubenswrapper[4983]: I0318 00:27:13.259089 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:27:13 crc kubenswrapper[4983]: I0318 00:27:13.259697 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:27:43 crc kubenswrapper[4983]: I0318 00:27:43.259319 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:27:43 crc kubenswrapper[4983]: I0318 00:27:43.259837 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:27:43 crc kubenswrapper[4983]: I0318 00:27:43.259887 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:27:43 crc kubenswrapper[4983]: I0318 00:27:43.260543 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f0dbf5791681286cf38e89c3a6323146d81246c217b7c32c7cf2e180e155fbe8"} pod="openshift-machine-config-operator/machine-config-daemon-jczfs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 00:27:43 crc kubenswrapper[4983]: I0318 00:27:43.260609 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" containerID="cri-o://f0dbf5791681286cf38e89c3a6323146d81246c217b7c32c7cf2e180e155fbe8" gracePeriod=600 Mar 18 00:27:44 crc kubenswrapper[4983]: I0318 00:27:44.004287 4983 generic.go:334] "Generic (PLEG): container finished" podID="482e2afe-48be-4ea2-8bda-52849c901b52" containerID="f0dbf5791681286cf38e89c3a6323146d81246c217b7c32c7cf2e180e155fbe8" exitCode=0 Mar 18 00:27:44 crc kubenswrapper[4983]: I0318 00:27:44.004393 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerDied","Data":"f0dbf5791681286cf38e89c3a6323146d81246c217b7c32c7cf2e180e155fbe8"} Mar 18 00:27:44 crc kubenswrapper[4983]: I0318 00:27:44.004694 4983 scope.go:117] "RemoveContainer" containerID="81d24af741ccda81f14aaade4f3fc2f7c8569f253f97a8c8f9b6fa245fbd1638" Mar 18 00:27:45 crc kubenswrapper[4983]: I0318 00:27:45.016282 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerStarted","Data":"daa639b5561e05c04168c0703cff853339f7b4e1eea5d3b72dde7a611b097532"} Mar 18 00:28:00 crc kubenswrapper[4983]: I0318 00:28:00.147794 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563228-bj5lk"] Mar 18 00:28:00 crc kubenswrapper[4983]: I0318 00:28:00.150006 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563228-bj5lk" Mar 18 00:28:00 crc kubenswrapper[4983]: I0318 00:28:00.153634 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:28:00 crc kubenswrapper[4983]: I0318 00:28:00.153642 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:28:00 crc kubenswrapper[4983]: I0318 00:28:00.156082 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:28:00 crc kubenswrapper[4983]: I0318 00:28:00.161891 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563228-bj5lk"] Mar 18 00:28:00 crc kubenswrapper[4983]: I0318 00:28:00.208952 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ccnj\" (UniqueName: \"kubernetes.io/projected/84ef092c-1aff-4381-8f9d-ec59958c1c1f-kube-api-access-8ccnj\") pod \"auto-csr-approver-29563228-bj5lk\" (UID: \"84ef092c-1aff-4381-8f9d-ec59958c1c1f\") " pod="openshift-infra/auto-csr-approver-29563228-bj5lk" Mar 18 00:28:00 crc kubenswrapper[4983]: I0318 00:28:00.310486 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ccnj\" (UniqueName: \"kubernetes.io/projected/84ef092c-1aff-4381-8f9d-ec59958c1c1f-kube-api-access-8ccnj\") pod \"auto-csr-approver-29563228-bj5lk\" (UID: \"84ef092c-1aff-4381-8f9d-ec59958c1c1f\") " pod="openshift-infra/auto-csr-approver-29563228-bj5lk" Mar 18 00:28:00 crc kubenswrapper[4983]: I0318 00:28:00.337686 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ccnj\" (UniqueName: \"kubernetes.io/projected/84ef092c-1aff-4381-8f9d-ec59958c1c1f-kube-api-access-8ccnj\") pod \"auto-csr-approver-29563228-bj5lk\" (UID: \"84ef092c-1aff-4381-8f9d-ec59958c1c1f\") " pod="openshift-infra/auto-csr-approver-29563228-bj5lk" Mar 18 00:28:00 crc kubenswrapper[4983]: I0318 00:28:00.479497 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563228-bj5lk" Mar 18 00:28:00 crc kubenswrapper[4983]: I0318 00:28:00.726062 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563228-bj5lk"] Mar 18 00:28:01 crc kubenswrapper[4983]: I0318 00:28:01.150956 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563228-bj5lk" event={"ID":"84ef092c-1aff-4381-8f9d-ec59958c1c1f","Type":"ContainerStarted","Data":"14dd2f27398fbc6e5c3d2f333cf9de9b5b54eba693aefcc43e172fc59f6d9024"} Mar 18 00:28:02 crc kubenswrapper[4983]: I0318 00:28:02.163001 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563228-bj5lk" event={"ID":"84ef092c-1aff-4381-8f9d-ec59958c1c1f","Type":"ContainerStarted","Data":"fd2afaab66d567abc373f78c2e812ec98a921ec7fb763c221086deb26161cfe1"} Mar 18 00:28:02 crc kubenswrapper[4983]: I0318 00:28:02.193113 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29563228-bj5lk" podStartSLOduration=1.172775156 podStartE2EDuration="2.193091162s" podCreationTimestamp="2026-03-18 00:28:00 +0000 UTC" firstStartedPulling="2026-03-18 00:28:00.732152425 +0000 UTC m=+1198.629527035" lastFinishedPulling="2026-03-18 00:28:01.752468401 +0000 UTC m=+1199.649843041" observedRunningTime="2026-03-18 00:28:02.184634524 +0000 UTC m=+1200.082009164" watchObservedRunningTime="2026-03-18 00:28:02.193091162 +0000 UTC m=+1200.090465792" Mar 18 00:28:03 crc kubenswrapper[4983]: I0318 00:28:03.172995 4983 generic.go:334] "Generic (PLEG): container finished" podID="84ef092c-1aff-4381-8f9d-ec59958c1c1f" containerID="fd2afaab66d567abc373f78c2e812ec98a921ec7fb763c221086deb26161cfe1" exitCode=0 Mar 18 00:28:03 crc kubenswrapper[4983]: I0318 00:28:03.173067 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563228-bj5lk" event={"ID":"84ef092c-1aff-4381-8f9d-ec59958c1c1f","Type":"ContainerDied","Data":"fd2afaab66d567abc373f78c2e812ec98a921ec7fb763c221086deb26161cfe1"} Mar 18 00:28:04 crc kubenswrapper[4983]: I0318 00:28:04.463659 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563228-bj5lk" Mar 18 00:28:04 crc kubenswrapper[4983]: I0318 00:28:04.508496 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ccnj\" (UniqueName: \"kubernetes.io/projected/84ef092c-1aff-4381-8f9d-ec59958c1c1f-kube-api-access-8ccnj\") pod \"84ef092c-1aff-4381-8f9d-ec59958c1c1f\" (UID: \"84ef092c-1aff-4381-8f9d-ec59958c1c1f\") " Mar 18 00:28:04 crc kubenswrapper[4983]: I0318 00:28:04.513642 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84ef092c-1aff-4381-8f9d-ec59958c1c1f-kube-api-access-8ccnj" (OuterVolumeSpecName: "kube-api-access-8ccnj") pod "84ef092c-1aff-4381-8f9d-ec59958c1c1f" (UID: "84ef092c-1aff-4381-8f9d-ec59958c1c1f"). InnerVolumeSpecName "kube-api-access-8ccnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:28:04 crc kubenswrapper[4983]: I0318 00:28:04.610327 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ccnj\" (UniqueName: \"kubernetes.io/projected/84ef092c-1aff-4381-8f9d-ec59958c1c1f-kube-api-access-8ccnj\") on node \"crc\" DevicePath \"\"" Mar 18 00:28:05 crc kubenswrapper[4983]: I0318 00:28:05.191914 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563228-bj5lk" event={"ID":"84ef092c-1aff-4381-8f9d-ec59958c1c1f","Type":"ContainerDied","Data":"14dd2f27398fbc6e5c3d2f333cf9de9b5b54eba693aefcc43e172fc59f6d9024"} Mar 18 00:28:05 crc kubenswrapper[4983]: I0318 00:28:05.191974 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14dd2f27398fbc6e5c3d2f333cf9de9b5b54eba693aefcc43e172fc59f6d9024" Mar 18 00:28:05 crc kubenswrapper[4983]: I0318 00:28:05.192017 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563228-bj5lk" Mar 18 00:28:05 crc kubenswrapper[4983]: I0318 00:28:05.256656 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563222-57gsb"] Mar 18 00:28:05 crc kubenswrapper[4983]: I0318 00:28:05.262423 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563222-57gsb"] Mar 18 00:28:06 crc kubenswrapper[4983]: I0318 00:28:06.858617 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c99a06e4-c46d-4036-8307-9eabfed2b5b0" path="/var/lib/kubelet/pods/c99a06e4-c46d-4036-8307-9eabfed2b5b0/volumes" Mar 18 00:29:10 crc kubenswrapper[4983]: I0318 00:29:10.082897 4983 scope.go:117] "RemoveContainer" containerID="6fb5448b2b229082f07dc8f36cd95fb67a21f88f4883fbe632c3dd0e769ecb02" Mar 18 00:29:29 crc kubenswrapper[4983]: I0318 00:29:29.842281 4983 generic.go:334] "Generic (PLEG): container finished" podID="e3f8b858-5134-424f-aa3f-fce238faead3" containerID="b8339df555734c7201081ab92dc2a7031ec373b4c37ad72fb7f1d9c0360f091e" exitCode=0 Mar 18 00:29:29 crc kubenswrapper[4983]: I0318 00:29:29.842407 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"e3f8b858-5134-424f-aa3f-fce238faead3","Type":"ContainerDied","Data":"b8339df555734c7201081ab92dc2a7031ec373b4c37ad72fb7f1d9c0360f091e"} Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.228036 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.363613 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e3f8b858-5134-424f-aa3f-fce238faead3-node-pullsecrets\") pod \"e3f8b858-5134-424f-aa3f-fce238faead3\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.363713 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/e3f8b858-5134-424f-aa3f-fce238faead3-builder-dockercfg-q2gzr-pull\") pod \"e3f8b858-5134-424f-aa3f-fce238faead3\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.363777 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e3f8b858-5134-424f-aa3f-fce238faead3-buildcachedir\") pod \"e3f8b858-5134-424f-aa3f-fce238faead3\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.363839 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-ca-bundles\") pod \"e3f8b858-5134-424f-aa3f-fce238faead3\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.364511 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3f8b858-5134-424f-aa3f-fce238faead3-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "e3f8b858-5134-424f-aa3f-fce238faead3" (UID: "e3f8b858-5134-424f-aa3f-fce238faead3"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.365652 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "e3f8b858-5134-424f-aa3f-fce238faead3" (UID: "e3f8b858-5134-424f-aa3f-fce238faead3"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.365727 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3f8b858-5134-424f-aa3f-fce238faead3-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "e3f8b858-5134-424f-aa3f-fce238faead3" (UID: "e3f8b858-5134-424f-aa3f-fce238faead3"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.366747 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "e3f8b858-5134-424f-aa3f-fce238faead3" (UID: "e3f8b858-5134-424f-aa3f-fce238faead3"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.368999 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-proxy-ca-bundles\") pod \"e3f8b858-5134-424f-aa3f-fce238faead3\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.369125 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-build-blob-cache\") pod \"e3f8b858-5134-424f-aa3f-fce238faead3\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.369203 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-container-storage-run\") pod \"e3f8b858-5134-424f-aa3f-fce238faead3\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.369264 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-buildworkdir\") pod \"e3f8b858-5134-424f-aa3f-fce238faead3\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.369336 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-system-configs\") pod \"e3f8b858-5134-424f-aa3f-fce238faead3\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.369437 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-container-storage-root\") pod \"e3f8b858-5134-424f-aa3f-fce238faead3\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.369508 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/e3f8b858-5134-424f-aa3f-fce238faead3-builder-dockercfg-q2gzr-push\") pod \"e3f8b858-5134-424f-aa3f-fce238faead3\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.369557 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9575w\" (UniqueName: \"kubernetes.io/projected/e3f8b858-5134-424f-aa3f-fce238faead3-kube-api-access-9575w\") pod \"e3f8b858-5134-424f-aa3f-fce238faead3\" (UID: \"e3f8b858-5134-424f-aa3f-fce238faead3\") " Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.370263 4983 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e3f8b858-5134-424f-aa3f-fce238faead3-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.370296 4983 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e3f8b858-5134-424f-aa3f-fce238faead3-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.370321 4983 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.370343 4983 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.377342 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3f8b858-5134-424f-aa3f-fce238faead3-kube-api-access-9575w" (OuterVolumeSpecName: "kube-api-access-9575w") pod "e3f8b858-5134-424f-aa3f-fce238faead3" (UID: "e3f8b858-5134-424f-aa3f-fce238faead3"). InnerVolumeSpecName "kube-api-access-9575w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.378983 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "e3f8b858-5134-424f-aa3f-fce238faead3" (UID: "e3f8b858-5134-424f-aa3f-fce238faead3"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.379806 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "e3f8b858-5134-424f-aa3f-fce238faead3" (UID: "e3f8b858-5134-424f-aa3f-fce238faead3"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.382133 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3f8b858-5134-424f-aa3f-fce238faead3-builder-dockercfg-q2gzr-push" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-push") pod "e3f8b858-5134-424f-aa3f-fce238faead3" (UID: "e3f8b858-5134-424f-aa3f-fce238faead3"). InnerVolumeSpecName "builder-dockercfg-q2gzr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.391997 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3f8b858-5134-424f-aa3f-fce238faead3-builder-dockercfg-q2gzr-pull" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-pull") pod "e3f8b858-5134-424f-aa3f-fce238faead3" (UID: "e3f8b858-5134-424f-aa3f-fce238faead3"). InnerVolumeSpecName "builder-dockercfg-q2gzr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.394242 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "e3f8b858-5134-424f-aa3f-fce238faead3" (UID: "e3f8b858-5134-424f-aa3f-fce238faead3"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.472373 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/e3f8b858-5134-424f-aa3f-fce238faead3-builder-dockercfg-q2gzr-push\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.472427 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9575w\" (UniqueName: \"kubernetes.io/projected/e3f8b858-5134-424f-aa3f-fce238faead3-kube-api-access-9575w\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.472450 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/e3f8b858-5134-424f-aa3f-fce238faead3-builder-dockercfg-q2gzr-pull\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.472471 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.472490 4983 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.472508 4983 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e3f8b858-5134-424f-aa3f-fce238faead3-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.725639 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "e3f8b858-5134-424f-aa3f-fce238faead3" (UID: "e3f8b858-5134-424f-aa3f-fce238faead3"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.776354 4983 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.864541 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"e3f8b858-5134-424f-aa3f-fce238faead3","Type":"ContainerDied","Data":"d26f5e0f286c0e6fa86a1261a0993d3353b2e2bd618fac4ca363bdc9ab680269"} Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.864583 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d26f5e0f286c0e6fa86a1261a0993d3353b2e2bd618fac4ca363bdc9ab680269" Mar 18 00:29:31 crc kubenswrapper[4983]: I0318 00:29:31.864640 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Mar 18 00:29:34 crc kubenswrapper[4983]: I0318 00:29:34.815141 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "e3f8b858-5134-424f-aa3f-fce238faead3" (UID: "e3f8b858-5134-424f-aa3f-fce238faead3"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:29:34 crc kubenswrapper[4983]: I0318 00:29:34.823523 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e3f8b858-5134-424f-aa3f-fce238faead3-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.283282 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 18 00:29:36 crc kubenswrapper[4983]: E0318 00:29:36.284546 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3f8b858-5134-424f-aa3f-fce238faead3" containerName="manage-dockerfile" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.284637 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3f8b858-5134-424f-aa3f-fce238faead3" containerName="manage-dockerfile" Mar 18 00:29:36 crc kubenswrapper[4983]: E0318 00:29:36.284723 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3f8b858-5134-424f-aa3f-fce238faead3" containerName="docker-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.284795 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3f8b858-5134-424f-aa3f-fce238faead3" containerName="docker-build" Mar 18 00:29:36 crc kubenswrapper[4983]: E0318 00:29:36.284872 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3f8b858-5134-424f-aa3f-fce238faead3" containerName="git-clone" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.284967 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3f8b858-5134-424f-aa3f-fce238faead3" containerName="git-clone" Mar 18 00:29:36 crc kubenswrapper[4983]: E0318 00:29:36.285056 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84ef092c-1aff-4381-8f9d-ec59958c1c1f" containerName="oc" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.285128 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="84ef092c-1aff-4381-8f9d-ec59958c1c1f" containerName="oc" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.285325 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="84ef092c-1aff-4381-8f9d-ec59958c1c1f" containerName="oc" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.285408 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3f8b858-5134-424f-aa3f-fce238faead3" containerName="docker-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.286230 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.291694 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-sys-config" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.292026 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-global-ca" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.292159 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-q2gzr" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.292065 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-ca" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.310445 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.444792 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/2894614e-f410-4ca5-86a5-1bc971c408be-builder-dockercfg-q2gzr-push\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.444852 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.444922 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2894614e-f410-4ca5-86a5-1bc971c408be-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.445019 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.445087 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j97vh\" (UniqueName: \"kubernetes.io/projected/2894614e-f410-4ca5-86a5-1bc971c408be-kube-api-access-j97vh\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.445256 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/2894614e-f410-4ca5-86a5-1bc971c408be-builder-dockercfg-q2gzr-pull\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.445343 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.445442 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.445648 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.445788 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2894614e-f410-4ca5-86a5-1bc971c408be-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.445876 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.445939 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.547791 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2894614e-f410-4ca5-86a5-1bc971c408be-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.547876 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.547963 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j97vh\" (UniqueName: \"kubernetes.io/projected/2894614e-f410-4ca5-86a5-1bc971c408be-kube-api-access-j97vh\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.548077 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/2894614e-f410-4ca5-86a5-1bc971c408be-builder-dockercfg-q2gzr-pull\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.548123 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.548138 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2894614e-f410-4ca5-86a5-1bc971c408be-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.548193 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.548285 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.548356 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2894614e-f410-4ca5-86a5-1bc971c408be-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.548417 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.548471 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.548548 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/2894614e-f410-4ca5-86a5-1bc971c408be-builder-dockercfg-q2gzr-push\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.548592 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.549372 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.549516 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.549626 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2894614e-f410-4ca5-86a5-1bc971c408be-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.550668 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.550806 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.551182 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.551791 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.554069 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.557536 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/2894614e-f410-4ca5-86a5-1bc971c408be-builder-dockercfg-q2gzr-pull\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.558045 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/2894614e-f410-4ca5-86a5-1bc971c408be-builder-dockercfg-q2gzr-push\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.585740 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j97vh\" (UniqueName: \"kubernetes.io/projected/2894614e-f410-4ca5-86a5-1bc971c408be-kube-api-access-j97vh\") pod \"sg-bridge-1-build\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.603302 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.822750 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 18 00:29:36 crc kubenswrapper[4983]: I0318 00:29:36.903749 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"2894614e-f410-4ca5-86a5-1bc971c408be","Type":"ContainerStarted","Data":"57d74927997c201c65aee8921ffccab23bb19a9e91cda738527b112770e3b868"} Mar 18 00:29:37 crc kubenswrapper[4983]: I0318 00:29:37.918143 4983 generic.go:334] "Generic (PLEG): container finished" podID="2894614e-f410-4ca5-86a5-1bc971c408be" containerID="95a25421ecfbc5e99611dd59bca2c51733480ce671baecaa388c63c7bd66f2f9" exitCode=0 Mar 18 00:29:37 crc kubenswrapper[4983]: I0318 00:29:37.918424 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"2894614e-f410-4ca5-86a5-1bc971c408be","Type":"ContainerDied","Data":"95a25421ecfbc5e99611dd59bca2c51733480ce671baecaa388c63c7bd66f2f9"} Mar 18 00:29:38 crc kubenswrapper[4983]: I0318 00:29:38.943095 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"2894614e-f410-4ca5-86a5-1bc971c408be","Type":"ContainerStarted","Data":"357bb5e02b3f6cd7bf528718ac53dc375ffbd704693214ba01e82a3482a9ae73"} Mar 18 00:29:38 crc kubenswrapper[4983]: I0318 00:29:38.982854 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-1-build" podStartSLOduration=2.982829291 podStartE2EDuration="2.982829291s" podCreationTimestamp="2026-03-18 00:29:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:29:38.97452303 +0000 UTC m=+1296.871897710" watchObservedRunningTime="2026-03-18 00:29:38.982829291 +0000 UTC m=+1296.880203911" Mar 18 00:29:46 crc kubenswrapper[4983]: I0318 00:29:46.634493 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 18 00:29:46 crc kubenswrapper[4983]: I0318 00:29:46.635423 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/sg-bridge-1-build" podUID="2894614e-f410-4ca5-86a5-1bc971c408be" containerName="docker-build" containerID="cri-o://357bb5e02b3f6cd7bf528718ac53dc375ffbd704693214ba01e82a3482a9ae73" gracePeriod=30 Mar 18 00:29:46 crc kubenswrapper[4983]: I0318 00:29:46.923796 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_2894614e-f410-4ca5-86a5-1bc971c408be/docker-build/0.log" Mar 18 00:29:46 crc kubenswrapper[4983]: I0318 00:29:46.924598 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.008733 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j97vh\" (UniqueName: \"kubernetes.io/projected/2894614e-f410-4ca5-86a5-1bc971c408be-kube-api-access-j97vh\") pod \"2894614e-f410-4ca5-86a5-1bc971c408be\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.009067 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2894614e-f410-4ca5-86a5-1bc971c408be-node-pullsecrets\") pod \"2894614e-f410-4ca5-86a5-1bc971c408be\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.009251 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-build-blob-cache\") pod \"2894614e-f410-4ca5-86a5-1bc971c408be\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.009431 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-system-configs\") pod \"2894614e-f410-4ca5-86a5-1bc971c408be\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.009593 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/2894614e-f410-4ca5-86a5-1bc971c408be-builder-dockercfg-q2gzr-pull\") pod \"2894614e-f410-4ca5-86a5-1bc971c408be\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.009784 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-ca-bundles\") pod \"2894614e-f410-4ca5-86a5-1bc971c408be\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.009144 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2894614e-f410-4ca5-86a5-1bc971c408be-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "2894614e-f410-4ca5-86a5-1bc971c408be" (UID: "2894614e-f410-4ca5-86a5-1bc971c408be"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.010040 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "2894614e-f410-4ca5-86a5-1bc971c408be" (UID: "2894614e-f410-4ca5-86a5-1bc971c408be"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.010148 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2894614e-f410-4ca5-86a5-1bc971c408be-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "2894614e-f410-4ca5-86a5-1bc971c408be" (UID: "2894614e-f410-4ca5-86a5-1bc971c408be"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.010858 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2894614e-f410-4ca5-86a5-1bc971c408be-buildcachedir\") pod \"2894614e-f410-4ca5-86a5-1bc971c408be\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.011029 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/2894614e-f410-4ca5-86a5-1bc971c408be-builder-dockercfg-q2gzr-push\") pod \"2894614e-f410-4ca5-86a5-1bc971c408be\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.011184 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-proxy-ca-bundles\") pod \"2894614e-f410-4ca5-86a5-1bc971c408be\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.012832 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "2894614e-f410-4ca5-86a5-1bc971c408be" (UID: "2894614e-f410-4ca5-86a5-1bc971c408be"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.011396 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-buildworkdir\") pod \"2894614e-f410-4ca5-86a5-1bc971c408be\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.013107 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-container-storage-root\") pod \"2894614e-f410-4ca5-86a5-1bc971c408be\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.013176 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-container-storage-run\") pod \"2894614e-f410-4ca5-86a5-1bc971c408be\" (UID: \"2894614e-f410-4ca5-86a5-1bc971c408be\") " Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.013671 4983 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2894614e-f410-4ca5-86a5-1bc971c408be-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.013690 4983 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.013700 4983 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.013709 4983 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2894614e-f410-4ca5-86a5-1bc971c408be-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.014249 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "2894614e-f410-4ca5-86a5-1bc971c408be" (UID: "2894614e-f410-4ca5-86a5-1bc971c408be"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.014622 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "2894614e-f410-4ca5-86a5-1bc971c408be" (UID: "2894614e-f410-4ca5-86a5-1bc971c408be"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.014849 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "2894614e-f410-4ca5-86a5-1bc971c408be" (UID: "2894614e-f410-4ca5-86a5-1bc971c408be"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.015487 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2894614e-f410-4ca5-86a5-1bc971c408be-kube-api-access-j97vh" (OuterVolumeSpecName: "kube-api-access-j97vh") pod "2894614e-f410-4ca5-86a5-1bc971c408be" (UID: "2894614e-f410-4ca5-86a5-1bc971c408be"). InnerVolumeSpecName "kube-api-access-j97vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.016462 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2894614e-f410-4ca5-86a5-1bc971c408be-builder-dockercfg-q2gzr-pull" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-pull") pod "2894614e-f410-4ca5-86a5-1bc971c408be" (UID: "2894614e-f410-4ca5-86a5-1bc971c408be"). InnerVolumeSpecName "builder-dockercfg-q2gzr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.018580 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2894614e-f410-4ca5-86a5-1bc971c408be-builder-dockercfg-q2gzr-push" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-push") pod "2894614e-f410-4ca5-86a5-1bc971c408be" (UID: "2894614e-f410-4ca5-86a5-1bc971c408be"). InnerVolumeSpecName "builder-dockercfg-q2gzr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.035144 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_2894614e-f410-4ca5-86a5-1bc971c408be/docker-build/0.log" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.036417 4983 generic.go:334] "Generic (PLEG): container finished" podID="2894614e-f410-4ca5-86a5-1bc971c408be" containerID="357bb5e02b3f6cd7bf528718ac53dc375ffbd704693214ba01e82a3482a9ae73" exitCode=1 Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.036514 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"2894614e-f410-4ca5-86a5-1bc971c408be","Type":"ContainerDied","Data":"357bb5e02b3f6cd7bf528718ac53dc375ffbd704693214ba01e82a3482a9ae73"} Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.036588 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"2894614e-f410-4ca5-86a5-1bc971c408be","Type":"ContainerDied","Data":"57d74927997c201c65aee8921ffccab23bb19a9e91cda738527b112770e3b868"} Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.036618 4983 scope.go:117] "RemoveContainer" containerID="357bb5e02b3f6cd7bf528718ac53dc375ffbd704693214ba01e82a3482a9ae73" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.036974 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.092158 4983 scope.go:117] "RemoveContainer" containerID="95a25421ecfbc5e99611dd59bca2c51733480ce671baecaa388c63c7bd66f2f9" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.114270 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "2894614e-f410-4ca5-86a5-1bc971c408be" (UID: "2894614e-f410-4ca5-86a5-1bc971c408be"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.115411 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/2894614e-f410-4ca5-86a5-1bc971c408be-builder-dockercfg-q2gzr-pull\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.115433 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/2894614e-f410-4ca5-86a5-1bc971c408be-builder-dockercfg-q2gzr-push\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.115444 4983 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2894614e-f410-4ca5-86a5-1bc971c408be-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.115456 4983 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.115470 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.115479 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j97vh\" (UniqueName: \"kubernetes.io/projected/2894614e-f410-4ca5-86a5-1bc971c408be-kube-api-access-j97vh\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.115488 4983 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.137319 4983 scope.go:117] "RemoveContainer" containerID="357bb5e02b3f6cd7bf528718ac53dc375ffbd704693214ba01e82a3482a9ae73" Mar 18 00:29:47 crc kubenswrapper[4983]: E0318 00:29:47.137994 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"357bb5e02b3f6cd7bf528718ac53dc375ffbd704693214ba01e82a3482a9ae73\": container with ID starting with 357bb5e02b3f6cd7bf528718ac53dc375ffbd704693214ba01e82a3482a9ae73 not found: ID does not exist" containerID="357bb5e02b3f6cd7bf528718ac53dc375ffbd704693214ba01e82a3482a9ae73" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.138038 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"357bb5e02b3f6cd7bf528718ac53dc375ffbd704693214ba01e82a3482a9ae73"} err="failed to get container status \"357bb5e02b3f6cd7bf528718ac53dc375ffbd704693214ba01e82a3482a9ae73\": rpc error: code = NotFound desc = could not find container \"357bb5e02b3f6cd7bf528718ac53dc375ffbd704693214ba01e82a3482a9ae73\": container with ID starting with 357bb5e02b3f6cd7bf528718ac53dc375ffbd704693214ba01e82a3482a9ae73 not found: ID does not exist" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.138067 4983 scope.go:117] "RemoveContainer" containerID="95a25421ecfbc5e99611dd59bca2c51733480ce671baecaa388c63c7bd66f2f9" Mar 18 00:29:47 crc kubenswrapper[4983]: E0318 00:29:47.138514 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95a25421ecfbc5e99611dd59bca2c51733480ce671baecaa388c63c7bd66f2f9\": container with ID starting with 95a25421ecfbc5e99611dd59bca2c51733480ce671baecaa388c63c7bd66f2f9 not found: ID does not exist" containerID="95a25421ecfbc5e99611dd59bca2c51733480ce671baecaa388c63c7bd66f2f9" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.138562 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95a25421ecfbc5e99611dd59bca2c51733480ce671baecaa388c63c7bd66f2f9"} err="failed to get container status \"95a25421ecfbc5e99611dd59bca2c51733480ce671baecaa388c63c7bd66f2f9\": rpc error: code = NotFound desc = could not find container \"95a25421ecfbc5e99611dd59bca2c51733480ce671baecaa388c63c7bd66f2f9\": container with ID starting with 95a25421ecfbc5e99611dd59bca2c51733480ce671baecaa388c63c7bd66f2f9 not found: ID does not exist" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.404169 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "2894614e-f410-4ca5-86a5-1bc971c408be" (UID: "2894614e-f410-4ca5-86a5-1bc971c408be"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.424152 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2894614e-f410-4ca5-86a5-1bc971c408be-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.713123 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 18 00:29:47 crc kubenswrapper[4983]: I0318 00:29:47.732828 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.282356 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-2-build"] Mar 18 00:29:48 crc kubenswrapper[4983]: E0318 00:29:48.283091 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2894614e-f410-4ca5-86a5-1bc971c408be" containerName="docker-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.283134 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2894614e-f410-4ca5-86a5-1bc971c408be" containerName="docker-build" Mar 18 00:29:48 crc kubenswrapper[4983]: E0318 00:29:48.283174 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2894614e-f410-4ca5-86a5-1bc971c408be" containerName="manage-dockerfile" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.283191 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2894614e-f410-4ca5-86a5-1bc971c408be" containerName="manage-dockerfile" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.283457 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="2894614e-f410-4ca5-86a5-1bc971c408be" containerName="docker-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.284822 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.288631 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-sys-config" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.288650 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-ca" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.288865 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-global-ca" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.289022 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-q2gzr" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.319667 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.349151 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.349209 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.349252 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2vdl\" (UniqueName: \"kubernetes.io/projected/603fd644-5a3a-4dd3-85f5-788c853516ea-kube-api-access-k2vdl\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.349278 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/603fd644-5a3a-4dd3-85f5-788c853516ea-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.349304 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.349350 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.349394 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.349432 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/603fd644-5a3a-4dd3-85f5-788c853516ea-builder-dockercfg-q2gzr-push\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.349460 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.349486 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.349520 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/603fd644-5a3a-4dd3-85f5-788c853516ea-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.349555 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/603fd644-5a3a-4dd3-85f5-788c853516ea-builder-dockercfg-q2gzr-pull\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.451729 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/603fd644-5a3a-4dd3-85f5-788c853516ea-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.451851 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/603fd644-5a3a-4dd3-85f5-788c853516ea-builder-dockercfg-q2gzr-pull\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.451915 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.452027 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.452109 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2vdl\" (UniqueName: \"kubernetes.io/projected/603fd644-5a3a-4dd3-85f5-788c853516ea-kube-api-access-k2vdl\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.452161 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/603fd644-5a3a-4dd3-85f5-788c853516ea-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.452210 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.452323 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.452383 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.452483 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/603fd644-5a3a-4dd3-85f5-788c853516ea-builder-dockercfg-q2gzr-push\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.452544 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.452593 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.453726 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/603fd644-5a3a-4dd3-85f5-788c853516ea-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.454050 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.454259 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.454330 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/603fd644-5a3a-4dd3-85f5-788c853516ea-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.454443 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.454577 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.454773 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.455137 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.455326 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.462588 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/603fd644-5a3a-4dd3-85f5-788c853516ea-builder-dockercfg-q2gzr-pull\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.463646 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/603fd644-5a3a-4dd3-85f5-788c853516ea-builder-dockercfg-q2gzr-push\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.480601 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2vdl\" (UniqueName: \"kubernetes.io/projected/603fd644-5a3a-4dd3-85f5-788c853516ea-kube-api-access-k2vdl\") pod \"sg-bridge-2-build\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.606374 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.860120 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2894614e-f410-4ca5-86a5-1bc971c408be" path="/var/lib/kubelet/pods/2894614e-f410-4ca5-86a5-1bc971c408be/volumes" Mar 18 00:29:48 crc kubenswrapper[4983]: I0318 00:29:48.921914 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Mar 18 00:29:49 crc kubenswrapper[4983]: I0318 00:29:49.064121 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"603fd644-5a3a-4dd3-85f5-788c853516ea","Type":"ContainerStarted","Data":"2678b5b4b099d16a769d39f2ab30ec7ea56c4f0ad63f701fe3b3b0787d6d8ec8"} Mar 18 00:29:50 crc kubenswrapper[4983]: I0318 00:29:50.073745 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"603fd644-5a3a-4dd3-85f5-788c853516ea","Type":"ContainerStarted","Data":"fe2d31a9ba96a9b1300df7b5cc2a3d37c3915b415f348c30c029bc0e3a32ccc1"} Mar 18 00:29:51 crc kubenswrapper[4983]: I0318 00:29:51.086194 4983 generic.go:334] "Generic (PLEG): container finished" podID="603fd644-5a3a-4dd3-85f5-788c853516ea" containerID="fe2d31a9ba96a9b1300df7b5cc2a3d37c3915b415f348c30c029bc0e3a32ccc1" exitCode=0 Mar 18 00:29:51 crc kubenswrapper[4983]: I0318 00:29:51.086480 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"603fd644-5a3a-4dd3-85f5-788c853516ea","Type":"ContainerDied","Data":"fe2d31a9ba96a9b1300df7b5cc2a3d37c3915b415f348c30c029bc0e3a32ccc1"} Mar 18 00:29:52 crc kubenswrapper[4983]: I0318 00:29:52.097509 4983 generic.go:334] "Generic (PLEG): container finished" podID="603fd644-5a3a-4dd3-85f5-788c853516ea" containerID="8abfde51502d94499ccea7f352e392b6e3a7ec014045ef3617d6111bb11c86a0" exitCode=0 Mar 18 00:29:52 crc kubenswrapper[4983]: I0318 00:29:52.097562 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"603fd644-5a3a-4dd3-85f5-788c853516ea","Type":"ContainerDied","Data":"8abfde51502d94499ccea7f352e392b6e3a7ec014045ef3617d6111bb11c86a0"} Mar 18 00:29:52 crc kubenswrapper[4983]: I0318 00:29:52.130809 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-2-build_603fd644-5a3a-4dd3-85f5-788c853516ea/manage-dockerfile/0.log" Mar 18 00:29:53 crc kubenswrapper[4983]: I0318 00:29:53.106024 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"603fd644-5a3a-4dd3-85f5-788c853516ea","Type":"ContainerStarted","Data":"17757612e66a47e8fa04a5933117974974d46b39e1581954f3845c7ee3ef1b0b"} Mar 18 00:29:53 crc kubenswrapper[4983]: I0318 00:29:53.149365 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-2-build" podStartSLOduration=5.149346302 podStartE2EDuration="5.149346302s" podCreationTimestamp="2026-03-18 00:29:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:29:53.147967133 +0000 UTC m=+1311.045341763" watchObservedRunningTime="2026-03-18 00:29:53.149346302 +0000 UTC m=+1311.046720912" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.150091 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563230-brcnh"] Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.151675 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563230-brcnh" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.155730 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.156079 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.158098 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.158127 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss"] Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.160072 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.164129 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.171205 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.178661 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563230-brcnh"] Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.188629 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss"] Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.224802 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6427533-b7a2-4f09-b180-528089d22e4e-config-volume\") pod \"collect-profiles-29563230-t9hss\" (UID: \"c6427533-b7a2-4f09-b180-528089d22e4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.224879 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb9pk\" (UniqueName: \"kubernetes.io/projected/c6427533-b7a2-4f09-b180-528089d22e4e-kube-api-access-sb9pk\") pod \"collect-profiles-29563230-t9hss\" (UID: \"c6427533-b7a2-4f09-b180-528089d22e4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.225002 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6427533-b7a2-4f09-b180-528089d22e4e-secret-volume\") pod \"collect-profiles-29563230-t9hss\" (UID: \"c6427533-b7a2-4f09-b180-528089d22e4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.225065 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wgxd\" (UniqueName: \"kubernetes.io/projected/dff6997b-f408-41b6-8541-4af59267837b-kube-api-access-9wgxd\") pod \"auto-csr-approver-29563230-brcnh\" (UID: \"dff6997b-f408-41b6-8541-4af59267837b\") " pod="openshift-infra/auto-csr-approver-29563230-brcnh" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.326326 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6427533-b7a2-4f09-b180-528089d22e4e-config-volume\") pod \"collect-profiles-29563230-t9hss\" (UID: \"c6427533-b7a2-4f09-b180-528089d22e4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.326897 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb9pk\" (UniqueName: \"kubernetes.io/projected/c6427533-b7a2-4f09-b180-528089d22e4e-kube-api-access-sb9pk\") pod \"collect-profiles-29563230-t9hss\" (UID: \"c6427533-b7a2-4f09-b180-528089d22e4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.327066 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6427533-b7a2-4f09-b180-528089d22e4e-secret-volume\") pod \"collect-profiles-29563230-t9hss\" (UID: \"c6427533-b7a2-4f09-b180-528089d22e4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.327626 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6427533-b7a2-4f09-b180-528089d22e4e-config-volume\") pod \"collect-profiles-29563230-t9hss\" (UID: \"c6427533-b7a2-4f09-b180-528089d22e4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.329306 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wgxd\" (UniqueName: \"kubernetes.io/projected/dff6997b-f408-41b6-8541-4af59267837b-kube-api-access-9wgxd\") pod \"auto-csr-approver-29563230-brcnh\" (UID: \"dff6997b-f408-41b6-8541-4af59267837b\") " pod="openshift-infra/auto-csr-approver-29563230-brcnh" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.343607 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6427533-b7a2-4f09-b180-528089d22e4e-secret-volume\") pod \"collect-profiles-29563230-t9hss\" (UID: \"c6427533-b7a2-4f09-b180-528089d22e4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.352734 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wgxd\" (UniqueName: \"kubernetes.io/projected/dff6997b-f408-41b6-8541-4af59267837b-kube-api-access-9wgxd\") pod \"auto-csr-approver-29563230-brcnh\" (UID: \"dff6997b-f408-41b6-8541-4af59267837b\") " pod="openshift-infra/auto-csr-approver-29563230-brcnh" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.359577 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb9pk\" (UniqueName: \"kubernetes.io/projected/c6427533-b7a2-4f09-b180-528089d22e4e-kube-api-access-sb9pk\") pod \"collect-profiles-29563230-t9hss\" (UID: \"c6427533-b7a2-4f09-b180-528089d22e4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.485979 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563230-brcnh" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.498404 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.880450 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss"] Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.889667 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563230-brcnh"] Mar 18 00:30:00 crc kubenswrapper[4983]: W0318 00:30:00.900115 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddff6997b_f408_41b6_8541_4af59267837b.slice/crio-11d14bf01ed044f605c11d60480674c1f77cde8ef3f0e81c6e7bea22a750e357 WatchSource:0}: Error finding container 11d14bf01ed044f605c11d60480674c1f77cde8ef3f0e81c6e7bea22a750e357: Status 404 returned error can't find the container with id 11d14bf01ed044f605c11d60480674c1f77cde8ef3f0e81c6e7bea22a750e357 Mar 18 00:30:00 crc kubenswrapper[4983]: I0318 00:30:00.902970 4983 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 00:30:01 crc kubenswrapper[4983]: I0318 00:30:01.181036 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563230-brcnh" event={"ID":"dff6997b-f408-41b6-8541-4af59267837b","Type":"ContainerStarted","Data":"11d14bf01ed044f605c11d60480674c1f77cde8ef3f0e81c6e7bea22a750e357"} Mar 18 00:30:01 crc kubenswrapper[4983]: I0318 00:30:01.183317 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" event={"ID":"c6427533-b7a2-4f09-b180-528089d22e4e","Type":"ContainerStarted","Data":"12e241fec4fc85ff57440502b8ab4c6428327ac3f8bdc794517e40b60e3f3753"} Mar 18 00:30:01 crc kubenswrapper[4983]: I0318 00:30:01.183799 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" event={"ID":"c6427533-b7a2-4f09-b180-528089d22e4e","Type":"ContainerStarted","Data":"781ff415541c6e5a0d985149198f02aa9085641c9576c05ed2c54035bd4f9da1"} Mar 18 00:30:01 crc kubenswrapper[4983]: I0318 00:30:01.219024 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" podStartSLOduration=1.219002846 podStartE2EDuration="1.219002846s" podCreationTimestamp="2026-03-18 00:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:30:01.213883133 +0000 UTC m=+1319.111257793" watchObservedRunningTime="2026-03-18 00:30:01.219002846 +0000 UTC m=+1319.116377466" Mar 18 00:30:02 crc kubenswrapper[4983]: I0318 00:30:02.197258 4983 generic.go:334] "Generic (PLEG): container finished" podID="c6427533-b7a2-4f09-b180-528089d22e4e" containerID="12e241fec4fc85ff57440502b8ab4c6428327ac3f8bdc794517e40b60e3f3753" exitCode=0 Mar 18 00:30:02 crc kubenswrapper[4983]: I0318 00:30:02.197330 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" event={"ID":"c6427533-b7a2-4f09-b180-528089d22e4e","Type":"ContainerDied","Data":"12e241fec4fc85ff57440502b8ab4c6428327ac3f8bdc794517e40b60e3f3753"} Mar 18 00:30:03 crc kubenswrapper[4983]: I0318 00:30:03.561249 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" Mar 18 00:30:03 crc kubenswrapper[4983]: I0318 00:30:03.677973 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb9pk\" (UniqueName: \"kubernetes.io/projected/c6427533-b7a2-4f09-b180-528089d22e4e-kube-api-access-sb9pk\") pod \"c6427533-b7a2-4f09-b180-528089d22e4e\" (UID: \"c6427533-b7a2-4f09-b180-528089d22e4e\") " Mar 18 00:30:03 crc kubenswrapper[4983]: I0318 00:30:03.678328 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6427533-b7a2-4f09-b180-528089d22e4e-secret-volume\") pod \"c6427533-b7a2-4f09-b180-528089d22e4e\" (UID: \"c6427533-b7a2-4f09-b180-528089d22e4e\") " Mar 18 00:30:03 crc kubenswrapper[4983]: I0318 00:30:03.678556 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6427533-b7a2-4f09-b180-528089d22e4e-config-volume\") pod \"c6427533-b7a2-4f09-b180-528089d22e4e\" (UID: \"c6427533-b7a2-4f09-b180-528089d22e4e\") " Mar 18 00:30:03 crc kubenswrapper[4983]: I0318 00:30:03.679870 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6427533-b7a2-4f09-b180-528089d22e4e-config-volume" (OuterVolumeSpecName: "config-volume") pod "c6427533-b7a2-4f09-b180-528089d22e4e" (UID: "c6427533-b7a2-4f09-b180-528089d22e4e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:30:03 crc kubenswrapper[4983]: I0318 00:30:03.686483 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6427533-b7a2-4f09-b180-528089d22e4e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c6427533-b7a2-4f09-b180-528089d22e4e" (UID: "c6427533-b7a2-4f09-b180-528089d22e4e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:30:03 crc kubenswrapper[4983]: I0318 00:30:03.686972 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6427533-b7a2-4f09-b180-528089d22e4e-kube-api-access-sb9pk" (OuterVolumeSpecName: "kube-api-access-sb9pk") pod "c6427533-b7a2-4f09-b180-528089d22e4e" (UID: "c6427533-b7a2-4f09-b180-528089d22e4e"). InnerVolumeSpecName "kube-api-access-sb9pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:30:03 crc kubenswrapper[4983]: I0318 00:30:03.780566 4983 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6427533-b7a2-4f09-b180-528089d22e4e-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:03 crc kubenswrapper[4983]: I0318 00:30:03.780967 4983 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6427533-b7a2-4f09-b180-528089d22e4e-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:03 crc kubenswrapper[4983]: I0318 00:30:03.781139 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb9pk\" (UniqueName: \"kubernetes.io/projected/c6427533-b7a2-4f09-b180-528089d22e4e-kube-api-access-sb9pk\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:04 crc kubenswrapper[4983]: I0318 00:30:04.223110 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" event={"ID":"c6427533-b7a2-4f09-b180-528089d22e4e","Type":"ContainerDied","Data":"781ff415541c6e5a0d985149198f02aa9085641c9576c05ed2c54035bd4f9da1"} Mar 18 00:30:04 crc kubenswrapper[4983]: I0318 00:30:04.223778 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="781ff415541c6e5a0d985149198f02aa9085641c9576c05ed2c54035bd4f9da1" Mar 18 00:30:04 crc kubenswrapper[4983]: I0318 00:30:04.223134 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29563230-t9hss" Mar 18 00:30:04 crc kubenswrapper[4983]: I0318 00:30:04.226431 4983 generic.go:334] "Generic (PLEG): container finished" podID="dff6997b-f408-41b6-8541-4af59267837b" containerID="766eeeacad8a1883b60a9c90cc628a1563d63b77977d3aa19a09de35cbe13275" exitCode=0 Mar 18 00:30:04 crc kubenswrapper[4983]: I0318 00:30:04.226507 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563230-brcnh" event={"ID":"dff6997b-f408-41b6-8541-4af59267837b","Type":"ContainerDied","Data":"766eeeacad8a1883b60a9c90cc628a1563d63b77977d3aa19a09de35cbe13275"} Mar 18 00:30:05 crc kubenswrapper[4983]: I0318 00:30:05.554107 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563230-brcnh" Mar 18 00:30:05 crc kubenswrapper[4983]: I0318 00:30:05.707137 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wgxd\" (UniqueName: \"kubernetes.io/projected/dff6997b-f408-41b6-8541-4af59267837b-kube-api-access-9wgxd\") pod \"dff6997b-f408-41b6-8541-4af59267837b\" (UID: \"dff6997b-f408-41b6-8541-4af59267837b\") " Mar 18 00:30:05 crc kubenswrapper[4983]: I0318 00:30:05.729384 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dff6997b-f408-41b6-8541-4af59267837b-kube-api-access-9wgxd" (OuterVolumeSpecName: "kube-api-access-9wgxd") pod "dff6997b-f408-41b6-8541-4af59267837b" (UID: "dff6997b-f408-41b6-8541-4af59267837b"). InnerVolumeSpecName "kube-api-access-9wgxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:30:05 crc kubenswrapper[4983]: I0318 00:30:05.808649 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wgxd\" (UniqueName: \"kubernetes.io/projected/dff6997b-f408-41b6-8541-4af59267837b-kube-api-access-9wgxd\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:06 crc kubenswrapper[4983]: I0318 00:30:06.244438 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563230-brcnh" event={"ID":"dff6997b-f408-41b6-8541-4af59267837b","Type":"ContainerDied","Data":"11d14bf01ed044f605c11d60480674c1f77cde8ef3f0e81c6e7bea22a750e357"} Mar 18 00:30:06 crc kubenswrapper[4983]: I0318 00:30:06.244886 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11d14bf01ed044f605c11d60480674c1f77cde8ef3f0e81c6e7bea22a750e357" Mar 18 00:30:06 crc kubenswrapper[4983]: I0318 00:30:06.244562 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563230-brcnh" Mar 18 00:30:06 crc kubenswrapper[4983]: I0318 00:30:06.617458 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563224-b4g22"] Mar 18 00:30:06 crc kubenswrapper[4983]: I0318 00:30:06.625720 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563224-b4g22"] Mar 18 00:30:06 crc kubenswrapper[4983]: I0318 00:30:06.861974 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15371e16-94d2-4fa9-a891-6168661fdaf7" path="/var/lib/kubelet/pods/15371e16-94d2-4fa9-a891-6168661fdaf7/volumes" Mar 18 00:30:10 crc kubenswrapper[4983]: I0318 00:30:10.183955 4983 scope.go:117] "RemoveContainer" containerID="e41b566259dca5a930a81be7a73d899985af9103788d9beb47f5b570ed6dfd52" Mar 18 00:30:13 crc kubenswrapper[4983]: I0318 00:30:13.259096 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:30:13 crc kubenswrapper[4983]: I0318 00:30:13.261239 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:30:42 crc kubenswrapper[4983]: I0318 00:30:42.582523 4983 generic.go:334] "Generic (PLEG): container finished" podID="603fd644-5a3a-4dd3-85f5-788c853516ea" containerID="17757612e66a47e8fa04a5933117974974d46b39e1581954f3845c7ee3ef1b0b" exitCode=0 Mar 18 00:30:42 crc kubenswrapper[4983]: I0318 00:30:42.583457 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"603fd644-5a3a-4dd3-85f5-788c853516ea","Type":"ContainerDied","Data":"17757612e66a47e8fa04a5933117974974d46b39e1581954f3845c7ee3ef1b0b"} Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.259115 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.259198 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.847017 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.992091 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-buildworkdir\") pod \"603fd644-5a3a-4dd3-85f5-788c853516ea\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.992180 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/603fd644-5a3a-4dd3-85f5-788c853516ea-builder-dockercfg-q2gzr-pull\") pod \"603fd644-5a3a-4dd3-85f5-788c853516ea\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.992220 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/603fd644-5a3a-4dd3-85f5-788c853516ea-buildcachedir\") pod \"603fd644-5a3a-4dd3-85f5-788c853516ea\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.992306 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-container-storage-root\") pod \"603fd644-5a3a-4dd3-85f5-788c853516ea\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.992359 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-proxy-ca-bundles\") pod \"603fd644-5a3a-4dd3-85f5-788c853516ea\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.992387 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/603fd644-5a3a-4dd3-85f5-788c853516ea-node-pullsecrets\") pod \"603fd644-5a3a-4dd3-85f5-788c853516ea\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.992387 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/603fd644-5a3a-4dd3-85f5-788c853516ea-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "603fd644-5a3a-4dd3-85f5-788c853516ea" (UID: "603fd644-5a3a-4dd3-85f5-788c853516ea"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.992427 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-system-configs\") pod \"603fd644-5a3a-4dd3-85f5-788c853516ea\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.992464 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/603fd644-5a3a-4dd3-85f5-788c853516ea-builder-dockercfg-q2gzr-push\") pod \"603fd644-5a3a-4dd3-85f5-788c853516ea\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.992514 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2vdl\" (UniqueName: \"kubernetes.io/projected/603fd644-5a3a-4dd3-85f5-788c853516ea-kube-api-access-k2vdl\") pod \"603fd644-5a3a-4dd3-85f5-788c853516ea\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.992557 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-build-blob-cache\") pod \"603fd644-5a3a-4dd3-85f5-788c853516ea\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.992611 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-container-storage-run\") pod \"603fd644-5a3a-4dd3-85f5-788c853516ea\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.992646 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-ca-bundles\") pod \"603fd644-5a3a-4dd3-85f5-788c853516ea\" (UID: \"603fd644-5a3a-4dd3-85f5-788c853516ea\") " Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.993126 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "603fd644-5a3a-4dd3-85f5-788c853516ea" (UID: "603fd644-5a3a-4dd3-85f5-788c853516ea"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.993183 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/603fd644-5a3a-4dd3-85f5-788c853516ea-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "603fd644-5a3a-4dd3-85f5-788c853516ea" (UID: "603fd644-5a3a-4dd3-85f5-788c853516ea"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.993299 4983 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/603fd644-5a3a-4dd3-85f5-788c853516ea-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.993324 4983 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/603fd644-5a3a-4dd3-85f5-788c853516ea-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.993343 4983 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.993834 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "603fd644-5a3a-4dd3-85f5-788c853516ea" (UID: "603fd644-5a3a-4dd3-85f5-788c853516ea"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.994305 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "603fd644-5a3a-4dd3-85f5-788c853516ea" (UID: "603fd644-5a3a-4dd3-85f5-788c853516ea"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.994761 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "603fd644-5a3a-4dd3-85f5-788c853516ea" (UID: "603fd644-5a3a-4dd3-85f5-788c853516ea"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.998418 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/603fd644-5a3a-4dd3-85f5-788c853516ea-builder-dockercfg-q2gzr-push" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-push") pod "603fd644-5a3a-4dd3-85f5-788c853516ea" (UID: "603fd644-5a3a-4dd3-85f5-788c853516ea"). InnerVolumeSpecName "builder-dockercfg-q2gzr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.998446 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "603fd644-5a3a-4dd3-85f5-788c853516ea" (UID: "603fd644-5a3a-4dd3-85f5-788c853516ea"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:30:43 crc kubenswrapper[4983]: I0318 00:30:43.999891 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/603fd644-5a3a-4dd3-85f5-788c853516ea-kube-api-access-k2vdl" (OuterVolumeSpecName: "kube-api-access-k2vdl") pod "603fd644-5a3a-4dd3-85f5-788c853516ea" (UID: "603fd644-5a3a-4dd3-85f5-788c853516ea"). InnerVolumeSpecName "kube-api-access-k2vdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.006570 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/603fd644-5a3a-4dd3-85f5-788c853516ea-builder-dockercfg-q2gzr-pull" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-pull") pod "603fd644-5a3a-4dd3-85f5-788c853516ea" (UID: "603fd644-5a3a-4dd3-85f5-788c853516ea"). InnerVolumeSpecName "builder-dockercfg-q2gzr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.094492 4983 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.094813 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/603fd644-5a3a-4dd3-85f5-788c853516ea-builder-dockercfg-q2gzr-pull\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.094976 4983 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.095126 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/603fd644-5a3a-4dd3-85f5-788c853516ea-builder-dockercfg-q2gzr-push\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.095250 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2vdl\" (UniqueName: \"kubernetes.io/projected/603fd644-5a3a-4dd3-85f5-788c853516ea-kube-api-access-k2vdl\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.095363 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.095480 4983 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/603fd644-5a3a-4dd3-85f5-788c853516ea-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.120378 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "603fd644-5a3a-4dd3-85f5-788c853516ea" (UID: "603fd644-5a3a-4dd3-85f5-788c853516ea"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.197336 4983 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.605788 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"603fd644-5a3a-4dd3-85f5-788c853516ea","Type":"ContainerDied","Data":"2678b5b4b099d16a769d39f2ab30ec7ea56c4f0ad63f701fe3b3b0787d6d8ec8"} Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.605845 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2678b5b4b099d16a769d39f2ab30ec7ea56c4f0ad63f701fe3b3b0787d6d8ec8" Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.605894 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Mar 18 00:30:44 crc kubenswrapper[4983]: I0318 00:30:44.984120 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "603fd644-5a3a-4dd3-85f5-788c853516ea" (UID: "603fd644-5a3a-4dd3-85f5-788c853516ea"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:30:45 crc kubenswrapper[4983]: I0318 00:30:45.012076 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/603fd644-5a3a-4dd3-85f5-788c853516ea-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.528257 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 18 00:30:48 crc kubenswrapper[4983]: E0318 00:30:48.530305 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dff6997b-f408-41b6-8541-4af59267837b" containerName="oc" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.530479 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="dff6997b-f408-41b6-8541-4af59267837b" containerName="oc" Mar 18 00:30:48 crc kubenswrapper[4983]: E0318 00:30:48.530653 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="603fd644-5a3a-4dd3-85f5-788c853516ea" containerName="manage-dockerfile" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.530821 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="603fd644-5a3a-4dd3-85f5-788c853516ea" containerName="manage-dockerfile" Mar 18 00:30:48 crc kubenswrapper[4983]: E0318 00:30:48.530998 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6427533-b7a2-4f09-b180-528089d22e4e" containerName="collect-profiles" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.531143 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6427533-b7a2-4f09-b180-528089d22e4e" containerName="collect-profiles" Mar 18 00:30:48 crc kubenswrapper[4983]: E0318 00:30:48.531339 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="603fd644-5a3a-4dd3-85f5-788c853516ea" containerName="git-clone" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.531467 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="603fd644-5a3a-4dd3-85f5-788c853516ea" containerName="git-clone" Mar 18 00:30:48 crc kubenswrapper[4983]: E0318 00:30:48.531597 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="603fd644-5a3a-4dd3-85f5-788c853516ea" containerName="docker-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.531722 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="603fd644-5a3a-4dd3-85f5-788c853516ea" containerName="docker-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.532180 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="603fd644-5a3a-4dd3-85f5-788c853516ea" containerName="docker-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.532345 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6427533-b7a2-4f09-b180-528089d22e4e" containerName="collect-profiles" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.532477 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="dff6997b-f408-41b6-8541-4af59267837b" containerName="oc" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.533687 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.537029 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-ca" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.537039 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-global-ca" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.540291 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-q2gzr" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.540360 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-sys-config" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.556468 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.673535 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/307f466e-6333-4615-a9ed-42f1b6be8c4f-builder-dockercfg-q2gzr-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.673594 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.673658 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzpgz\" (UniqueName: \"kubernetes.io/projected/307f466e-6333-4615-a9ed-42f1b6be8c4f-kube-api-access-lzpgz\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.674108 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.674167 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.674358 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.674409 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/307f466e-6333-4615-a9ed-42f1b6be8c4f-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.674435 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/307f466e-6333-4615-a9ed-42f1b6be8c4f-builder-dockercfg-q2gzr-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.674531 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.674562 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/307f466e-6333-4615-a9ed-42f1b6be8c4f-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.674629 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.674760 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.775412 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.775785 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.775804 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzpgz\" (UniqueName: \"kubernetes.io/projected/307f466e-6333-4615-a9ed-42f1b6be8c4f-kube-api-access-lzpgz\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.775823 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.775837 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/307f466e-6333-4615-a9ed-42f1b6be8c4f-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.775854 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/307f466e-6333-4615-a9ed-42f1b6be8c4f-builder-dockercfg-q2gzr-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.775885 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.775902 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/307f466e-6333-4615-a9ed-42f1b6be8c4f-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.775921 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.775975 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.776015 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/307f466e-6333-4615-a9ed-42f1b6be8c4f-builder-dockercfg-q2gzr-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.776037 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.776425 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.776754 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.776815 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.776848 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/307f466e-6333-4615-a9ed-42f1b6be8c4f-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.777054 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.777055 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.777638 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.777687 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.777722 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/307f466e-6333-4615-a9ed-42f1b6be8c4f-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.786950 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/307f466e-6333-4615-a9ed-42f1b6be8c4f-builder-dockercfg-q2gzr-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.787004 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/307f466e-6333-4615-a9ed-42f1b6be8c4f-builder-dockercfg-q2gzr-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.794295 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzpgz\" (UniqueName: \"kubernetes.io/projected/307f466e-6333-4615-a9ed-42f1b6be8c4f-kube-api-access-lzpgz\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:48 crc kubenswrapper[4983]: I0318 00:30:48.851457 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:49 crc kubenswrapper[4983]: I0318 00:30:49.094650 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 18 00:30:49 crc kubenswrapper[4983]: I0318 00:30:49.654208 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"307f466e-6333-4615-a9ed-42f1b6be8c4f","Type":"ContainerStarted","Data":"30d51b2926a6aa18901dbb1da2acc1d758ea4637c9693e4dd173beaf07d6499a"} Mar 18 00:30:49 crc kubenswrapper[4983]: I0318 00:30:49.654287 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"307f466e-6333-4615-a9ed-42f1b6be8c4f","Type":"ContainerStarted","Data":"7f2a4e67bee3d866d5dcceb8fbe2c3365fac09a1076fef21f1d8fbec5917339b"} Mar 18 00:30:50 crc kubenswrapper[4983]: I0318 00:30:50.667638 4983 generic.go:334] "Generic (PLEG): container finished" podID="307f466e-6333-4615-a9ed-42f1b6be8c4f" containerID="30d51b2926a6aa18901dbb1da2acc1d758ea4637c9693e4dd173beaf07d6499a" exitCode=0 Mar 18 00:30:50 crc kubenswrapper[4983]: I0318 00:30:50.667744 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"307f466e-6333-4615-a9ed-42f1b6be8c4f","Type":"ContainerDied","Data":"30d51b2926a6aa18901dbb1da2acc1d758ea4637c9693e4dd173beaf07d6499a"} Mar 18 00:30:50 crc kubenswrapper[4983]: I0318 00:30:50.668129 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"307f466e-6333-4615-a9ed-42f1b6be8c4f","Type":"ContainerStarted","Data":"5723b08347005e3df6b009ca77ab3f10fd1697520633978fc54f7ce71faec70c"} Mar 18 00:30:50 crc kubenswrapper[4983]: I0318 00:30:50.705910 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-1-build" podStartSLOduration=2.7058852 podStartE2EDuration="2.7058852s" podCreationTimestamp="2026-03-18 00:30:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:30:50.699391259 +0000 UTC m=+1368.596765899" watchObservedRunningTime="2026-03-18 00:30:50.7058852 +0000 UTC m=+1368.603259840" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.261599 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.263010 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/prometheus-webhook-snmp-1-build" podUID="307f466e-6333-4615-a9ed-42f1b6be8c4f" containerName="docker-build" containerID="cri-o://5723b08347005e3df6b009ca77ab3f10fd1697520633978fc54f7ce71faec70c" gracePeriod=30 Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.649174 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_307f466e-6333-4615-a9ed-42f1b6be8c4f/docker-build/0.log" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.650000 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.746357 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_307f466e-6333-4615-a9ed-42f1b6be8c4f/docker-build/0.log" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.747007 4983 generic.go:334] "Generic (PLEG): container finished" podID="307f466e-6333-4615-a9ed-42f1b6be8c4f" containerID="5723b08347005e3df6b009ca77ab3f10fd1697520633978fc54f7ce71faec70c" exitCode=1 Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.747074 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"307f466e-6333-4615-a9ed-42f1b6be8c4f","Type":"ContainerDied","Data":"5723b08347005e3df6b009ca77ab3f10fd1697520633978fc54f7ce71faec70c"} Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.747118 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.747162 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"307f466e-6333-4615-a9ed-42f1b6be8c4f","Type":"ContainerDied","Data":"7f2a4e67bee3d866d5dcceb8fbe2c3365fac09a1076fef21f1d8fbec5917339b"} Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.747201 4983 scope.go:117] "RemoveContainer" containerID="5723b08347005e3df6b009ca77ab3f10fd1697520633978fc54f7ce71faec70c" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.780573 4983 scope.go:117] "RemoveContainer" containerID="30d51b2926a6aa18901dbb1da2acc1d758ea4637c9693e4dd173beaf07d6499a" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.818610 4983 scope.go:117] "RemoveContainer" containerID="5723b08347005e3df6b009ca77ab3f10fd1697520633978fc54f7ce71faec70c" Mar 18 00:30:59 crc kubenswrapper[4983]: E0318 00:30:59.819196 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5723b08347005e3df6b009ca77ab3f10fd1697520633978fc54f7ce71faec70c\": container with ID starting with 5723b08347005e3df6b009ca77ab3f10fd1697520633978fc54f7ce71faec70c not found: ID does not exist" containerID="5723b08347005e3df6b009ca77ab3f10fd1697520633978fc54f7ce71faec70c" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.819250 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5723b08347005e3df6b009ca77ab3f10fd1697520633978fc54f7ce71faec70c"} err="failed to get container status \"5723b08347005e3df6b009ca77ab3f10fd1697520633978fc54f7ce71faec70c\": rpc error: code = NotFound desc = could not find container \"5723b08347005e3df6b009ca77ab3f10fd1697520633978fc54f7ce71faec70c\": container with ID starting with 5723b08347005e3df6b009ca77ab3f10fd1697520633978fc54f7ce71faec70c not found: ID does not exist" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.819284 4983 scope.go:117] "RemoveContainer" containerID="30d51b2926a6aa18901dbb1da2acc1d758ea4637c9693e4dd173beaf07d6499a" Mar 18 00:30:59 crc kubenswrapper[4983]: E0318 00:30:59.820002 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30d51b2926a6aa18901dbb1da2acc1d758ea4637c9693e4dd173beaf07d6499a\": container with ID starting with 30d51b2926a6aa18901dbb1da2acc1d758ea4637c9693e4dd173beaf07d6499a not found: ID does not exist" containerID="30d51b2926a6aa18901dbb1da2acc1d758ea4637c9693e4dd173beaf07d6499a" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.820035 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30d51b2926a6aa18901dbb1da2acc1d758ea4637c9693e4dd173beaf07d6499a"} err="failed to get container status \"30d51b2926a6aa18901dbb1da2acc1d758ea4637c9693e4dd173beaf07d6499a\": rpc error: code = NotFound desc = could not find container \"30d51b2926a6aa18901dbb1da2acc1d758ea4637c9693e4dd173beaf07d6499a\": container with ID starting with 30d51b2926a6aa18901dbb1da2acc1d758ea4637c9693e4dd173beaf07d6499a not found: ID does not exist" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.846765 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-container-storage-root\") pod \"307f466e-6333-4615-a9ed-42f1b6be8c4f\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.846866 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/307f466e-6333-4615-a9ed-42f1b6be8c4f-buildcachedir\") pod \"307f466e-6333-4615-a9ed-42f1b6be8c4f\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847014 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-proxy-ca-bundles\") pod \"307f466e-6333-4615-a9ed-42f1b6be8c4f\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847034 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/307f466e-6333-4615-a9ed-42f1b6be8c4f-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "307f466e-6333-4615-a9ed-42f1b6be8c4f" (UID: "307f466e-6333-4615-a9ed-42f1b6be8c4f"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847108 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-system-configs\") pod \"307f466e-6333-4615-a9ed-42f1b6be8c4f\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847164 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/307f466e-6333-4615-a9ed-42f1b6be8c4f-builder-dockercfg-q2gzr-pull\") pod \"307f466e-6333-4615-a9ed-42f1b6be8c4f\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847225 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzpgz\" (UniqueName: \"kubernetes.io/projected/307f466e-6333-4615-a9ed-42f1b6be8c4f-kube-api-access-lzpgz\") pod \"307f466e-6333-4615-a9ed-42f1b6be8c4f\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847274 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/307f466e-6333-4615-a9ed-42f1b6be8c4f-builder-dockercfg-q2gzr-push\") pod \"307f466e-6333-4615-a9ed-42f1b6be8c4f\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847327 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-buildworkdir\") pod \"307f466e-6333-4615-a9ed-42f1b6be8c4f\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847384 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-ca-bundles\") pod \"307f466e-6333-4615-a9ed-42f1b6be8c4f\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847468 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/307f466e-6333-4615-a9ed-42f1b6be8c4f-node-pullsecrets\") pod \"307f466e-6333-4615-a9ed-42f1b6be8c4f\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847554 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-blob-cache\") pod \"307f466e-6333-4615-a9ed-42f1b6be8c4f\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847603 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-container-storage-run\") pod \"307f466e-6333-4615-a9ed-42f1b6be8c4f\" (UID: \"307f466e-6333-4615-a9ed-42f1b6be8c4f\") " Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847738 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/307f466e-6333-4615-a9ed-42f1b6be8c4f-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "307f466e-6333-4615-a9ed-42f1b6be8c4f" (UID: "307f466e-6333-4615-a9ed-42f1b6be8c4f"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847964 4983 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/307f466e-6333-4615-a9ed-42f1b6be8c4f-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.847987 4983 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/307f466e-6333-4615-a9ed-42f1b6be8c4f-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.848359 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "307f466e-6333-4615-a9ed-42f1b6be8c4f" (UID: "307f466e-6333-4615-a9ed-42f1b6be8c4f"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.848650 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "307f466e-6333-4615-a9ed-42f1b6be8c4f" (UID: "307f466e-6333-4615-a9ed-42f1b6be8c4f"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.848729 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "307f466e-6333-4615-a9ed-42f1b6be8c4f" (UID: "307f466e-6333-4615-a9ed-42f1b6be8c4f"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.848766 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "307f466e-6333-4615-a9ed-42f1b6be8c4f" (UID: "307f466e-6333-4615-a9ed-42f1b6be8c4f"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.849807 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "307f466e-6333-4615-a9ed-42f1b6be8c4f" (UID: "307f466e-6333-4615-a9ed-42f1b6be8c4f"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.855574 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/307f466e-6333-4615-a9ed-42f1b6be8c4f-builder-dockercfg-q2gzr-push" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-push") pod "307f466e-6333-4615-a9ed-42f1b6be8c4f" (UID: "307f466e-6333-4615-a9ed-42f1b6be8c4f"). InnerVolumeSpecName "builder-dockercfg-q2gzr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.855784 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/307f466e-6333-4615-a9ed-42f1b6be8c4f-builder-dockercfg-q2gzr-pull" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-pull") pod "307f466e-6333-4615-a9ed-42f1b6be8c4f" (UID: "307f466e-6333-4615-a9ed-42f1b6be8c4f"). InnerVolumeSpecName "builder-dockercfg-q2gzr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.857469 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/307f466e-6333-4615-a9ed-42f1b6be8c4f-kube-api-access-lzpgz" (OuterVolumeSpecName: "kube-api-access-lzpgz") pod "307f466e-6333-4615-a9ed-42f1b6be8c4f" (UID: "307f466e-6333-4615-a9ed-42f1b6be8c4f"). InnerVolumeSpecName "kube-api-access-lzpgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.928267 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "307f466e-6333-4615-a9ed-42f1b6be8c4f" (UID: "307f466e-6333-4615-a9ed-42f1b6be8c4f"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.948790 4983 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.948834 4983 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.948853 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/307f466e-6333-4615-a9ed-42f1b6be8c4f-builder-dockercfg-q2gzr-pull\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.948872 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzpgz\" (UniqueName: \"kubernetes.io/projected/307f466e-6333-4615-a9ed-42f1b6be8c4f-kube-api-access-lzpgz\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.948888 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/307f466e-6333-4615-a9ed-42f1b6be8c4f-builder-dockercfg-q2gzr-push\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.948907 4983 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.948924 4983 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.948967 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 18 00:30:59 crc kubenswrapper[4983]: I0318 00:30:59.948983 4983 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.433153 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "307f466e-6333-4615-a9ed-42f1b6be8c4f" (UID: "307f466e-6333-4615-a9ed-42f1b6be8c4f"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.454448 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/307f466e-6333-4615-a9ed-42f1b6be8c4f-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.711990 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.721479 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.857170 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="307f466e-6333-4615-a9ed-42f1b6be8c4f" path="/var/lib/kubelet/pods/307f466e-6333-4615-a9ed-42f1b6be8c4f/volumes" Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.904703 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Mar 18 00:31:00 crc kubenswrapper[4983]: E0318 00:31:00.905194 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="307f466e-6333-4615-a9ed-42f1b6be8c4f" containerName="docker-build" Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.905229 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="307f466e-6333-4615-a9ed-42f1b6be8c4f" containerName="docker-build" Mar 18 00:31:00 crc kubenswrapper[4983]: E0318 00:31:00.905251 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="307f466e-6333-4615-a9ed-42f1b6be8c4f" containerName="manage-dockerfile" Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.905267 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="307f466e-6333-4615-a9ed-42f1b6be8c4f" containerName="manage-dockerfile" Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.905553 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="307f466e-6333-4615-a9ed-42f1b6be8c4f" containerName="docker-build" Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.907278 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.911162 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-global-ca" Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.911294 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-ca" Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.912252 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-sys-config" Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.912339 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-q2gzr" Mar 18 00:31:00 crc kubenswrapper[4983]: I0318 00:31:00.945323 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.074110 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.074201 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-builder-dockercfg-q2gzr-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.074630 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbx9r\" (UniqueName: \"kubernetes.io/projected/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-kube-api-access-fbx9r\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.074728 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-builder-dockercfg-q2gzr-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.074845 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.074920 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.075078 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.075124 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.075163 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.075237 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.075280 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.075312 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.176731 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-builder-dockercfg-q2gzr-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.176869 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-builder-dockercfg-q2gzr-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.176909 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbx9r\" (UniqueName: \"kubernetes.io/projected/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-kube-api-access-fbx9r\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.176980 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.177023 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.177087 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.177095 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.177130 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.177264 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.177344 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.177415 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.177460 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.177528 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.177624 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.178030 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.178095 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.178338 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.178460 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.179575 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.179957 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.181379 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.186972 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-builder-dockercfg-q2gzr-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.186995 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-builder-dockercfg-q2gzr-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.207788 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbx9r\" (UniqueName: \"kubernetes.io/projected/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-kube-api-access-fbx9r\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.232300 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.519721 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Mar 18 00:31:01 crc kubenswrapper[4983]: I0318 00:31:01.769653 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"51f72883-b55f-4b66-a0e3-36d5cb3ceebc","Type":"ContainerStarted","Data":"79b8c5f4ae1c38cecd1d6cf1205b62c85bbbf1ebe1d71717d8b873ffaeb710ae"} Mar 18 00:31:02 crc kubenswrapper[4983]: I0318 00:31:02.779292 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"51f72883-b55f-4b66-a0e3-36d5cb3ceebc","Type":"ContainerStarted","Data":"dc543649566138cfc7bab17d2ef072f04240415fd5de69f665debc2965580bac"} Mar 18 00:31:03 crc kubenswrapper[4983]: E0318 00:31:03.081749 4983 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.217:58600->38.102.83.217:43207: write tcp 38.102.83.217:58600->38.102.83.217:43207: write: broken pipe Mar 18 00:31:03 crc kubenswrapper[4983]: I0318 00:31:03.791399 4983 generic.go:334] "Generic (PLEG): container finished" podID="51f72883-b55f-4b66-a0e3-36d5cb3ceebc" containerID="dc543649566138cfc7bab17d2ef072f04240415fd5de69f665debc2965580bac" exitCode=0 Mar 18 00:31:03 crc kubenswrapper[4983]: I0318 00:31:03.791506 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"51f72883-b55f-4b66-a0e3-36d5cb3ceebc","Type":"ContainerDied","Data":"dc543649566138cfc7bab17d2ef072f04240415fd5de69f665debc2965580bac"} Mar 18 00:31:04 crc kubenswrapper[4983]: I0318 00:31:04.803528 4983 generic.go:334] "Generic (PLEG): container finished" podID="51f72883-b55f-4b66-a0e3-36d5cb3ceebc" containerID="589f587589af8b7b9c830f7388dcc3791a71ef79e07bc41b51febb69adc497f0" exitCode=0 Mar 18 00:31:04 crc kubenswrapper[4983]: I0318 00:31:04.803581 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"51f72883-b55f-4b66-a0e3-36d5cb3ceebc","Type":"ContainerDied","Data":"589f587589af8b7b9c830f7388dcc3791a71ef79e07bc41b51febb69adc497f0"} Mar 18 00:31:04 crc kubenswrapper[4983]: I0318 00:31:04.859724 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-2-build_51f72883-b55f-4b66-a0e3-36d5cb3ceebc/manage-dockerfile/0.log" Mar 18 00:31:05 crc kubenswrapper[4983]: I0318 00:31:05.815647 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"51f72883-b55f-4b66-a0e3-36d5cb3ceebc","Type":"ContainerStarted","Data":"f85b5d6869e3f701d7184f4c42c3a642e9fbde5ed69fd58b4ec5f1f70a30a887"} Mar 18 00:31:05 crc kubenswrapper[4983]: I0318 00:31:05.885191 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-2-build" podStartSLOduration=5.88515998 podStartE2EDuration="5.88515998s" podCreationTimestamp="2026-03-18 00:31:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:31:05.876617802 +0000 UTC m=+1383.773992462" watchObservedRunningTime="2026-03-18 00:31:05.88515998 +0000 UTC m=+1383.782534630" Mar 18 00:31:13 crc kubenswrapper[4983]: I0318 00:31:13.260114 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:31:13 crc kubenswrapper[4983]: I0318 00:31:13.261528 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:31:13 crc kubenswrapper[4983]: I0318 00:31:13.261769 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:31:13 crc kubenswrapper[4983]: I0318 00:31:13.262284 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"daa639b5561e05c04168c0703cff853339f7b4e1eea5d3b72dde7a611b097532"} pod="openshift-machine-config-operator/machine-config-daemon-jczfs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 00:31:13 crc kubenswrapper[4983]: I0318 00:31:13.262329 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" containerID="cri-o://daa639b5561e05c04168c0703cff853339f7b4e1eea5d3b72dde7a611b097532" gracePeriod=600 Mar 18 00:31:13 crc kubenswrapper[4983]: I0318 00:31:13.894320 4983 generic.go:334] "Generic (PLEG): container finished" podID="482e2afe-48be-4ea2-8bda-52849c901b52" containerID="daa639b5561e05c04168c0703cff853339f7b4e1eea5d3b72dde7a611b097532" exitCode=0 Mar 18 00:31:13 crc kubenswrapper[4983]: I0318 00:31:13.894358 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerDied","Data":"daa639b5561e05c04168c0703cff853339f7b4e1eea5d3b72dde7a611b097532"} Mar 18 00:31:13 crc kubenswrapper[4983]: I0318 00:31:13.894920 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerStarted","Data":"fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9"} Mar 18 00:31:13 crc kubenswrapper[4983]: I0318 00:31:13.894963 4983 scope.go:117] "RemoveContainer" containerID="f0dbf5791681286cf38e89c3a6323146d81246c217b7c32c7cf2e180e155fbe8" Mar 18 00:32:00 crc kubenswrapper[4983]: I0318 00:32:00.168681 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563232-jl6dt"] Mar 18 00:32:00 crc kubenswrapper[4983]: I0318 00:32:00.170636 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563232-jl6dt" Mar 18 00:32:00 crc kubenswrapper[4983]: I0318 00:32:00.174679 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:32:00 crc kubenswrapper[4983]: I0318 00:32:00.174810 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:32:00 crc kubenswrapper[4983]: I0318 00:32:00.176081 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:32:00 crc kubenswrapper[4983]: I0318 00:32:00.179188 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563232-jl6dt"] Mar 18 00:32:00 crc kubenswrapper[4983]: I0318 00:32:00.254446 4983 generic.go:334] "Generic (PLEG): container finished" podID="51f72883-b55f-4b66-a0e3-36d5cb3ceebc" containerID="f85b5d6869e3f701d7184f4c42c3a642e9fbde5ed69fd58b4ec5f1f70a30a887" exitCode=0 Mar 18 00:32:00 crc kubenswrapper[4983]: I0318 00:32:00.254588 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"51f72883-b55f-4b66-a0e3-36d5cb3ceebc","Type":"ContainerDied","Data":"f85b5d6869e3f701d7184f4c42c3a642e9fbde5ed69fd58b4ec5f1f70a30a887"} Mar 18 00:32:00 crc kubenswrapper[4983]: I0318 00:32:00.299521 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w85cb\" (UniqueName: \"kubernetes.io/projected/5db34645-4368-4df2-9da6-f3b4b119a4cf-kube-api-access-w85cb\") pod \"auto-csr-approver-29563232-jl6dt\" (UID: \"5db34645-4368-4df2-9da6-f3b4b119a4cf\") " pod="openshift-infra/auto-csr-approver-29563232-jl6dt" Mar 18 00:32:00 crc kubenswrapper[4983]: I0318 00:32:00.401120 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w85cb\" (UniqueName: \"kubernetes.io/projected/5db34645-4368-4df2-9da6-f3b4b119a4cf-kube-api-access-w85cb\") pod \"auto-csr-approver-29563232-jl6dt\" (UID: \"5db34645-4368-4df2-9da6-f3b4b119a4cf\") " pod="openshift-infra/auto-csr-approver-29563232-jl6dt" Mar 18 00:32:00 crc kubenswrapper[4983]: I0318 00:32:00.525612 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w85cb\" (UniqueName: \"kubernetes.io/projected/5db34645-4368-4df2-9da6-f3b4b119a4cf-kube-api-access-w85cb\") pod \"auto-csr-approver-29563232-jl6dt\" (UID: \"5db34645-4368-4df2-9da6-f3b4b119a4cf\") " pod="openshift-infra/auto-csr-approver-29563232-jl6dt" Mar 18 00:32:00 crc kubenswrapper[4983]: I0318 00:32:00.533383 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563232-jl6dt" Mar 18 00:32:00 crc kubenswrapper[4983]: I0318 00:32:00.780151 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563232-jl6dt"] Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.266831 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563232-jl6dt" event={"ID":"5db34645-4368-4df2-9da6-f3b4b119a4cf","Type":"ContainerStarted","Data":"4e805c09efc5ff12da49751d8aa211ff59765f8dc251ac1966d668e21b753c9b"} Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.677730 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.820258 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-blob-cache\") pod \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.820341 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-ca-bundles\") pod \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.820391 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-container-storage-root\") pod \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.820432 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-proxy-ca-bundles\") pod \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.820464 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-node-pullsecrets\") pod \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.820520 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbx9r\" (UniqueName: \"kubernetes.io/projected/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-kube-api-access-fbx9r\") pod \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.820559 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-container-storage-run\") pod \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.820590 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-buildcachedir\") pod \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.820637 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-builder-dockercfg-q2gzr-pull\") pod \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.820736 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-buildworkdir\") pod \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.820792 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-builder-dockercfg-q2gzr-push\") pod \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.820861 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-system-configs\") pod \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\" (UID: \"51f72883-b55f-4b66-a0e3-36d5cb3ceebc\") " Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.821898 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "51f72883-b55f-4b66-a0e3-36d5cb3ceebc" (UID: "51f72883-b55f-4b66-a0e3-36d5cb3ceebc"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.822110 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "51f72883-b55f-4b66-a0e3-36d5cb3ceebc" (UID: "51f72883-b55f-4b66-a0e3-36d5cb3ceebc"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.822234 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "51f72883-b55f-4b66-a0e3-36d5cb3ceebc" (UID: "51f72883-b55f-4b66-a0e3-36d5cb3ceebc"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.822274 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "51f72883-b55f-4b66-a0e3-36d5cb3ceebc" (UID: "51f72883-b55f-4b66-a0e3-36d5cb3ceebc"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.822830 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "51f72883-b55f-4b66-a0e3-36d5cb3ceebc" (UID: "51f72883-b55f-4b66-a0e3-36d5cb3ceebc"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.823253 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "51f72883-b55f-4b66-a0e3-36d5cb3ceebc" (UID: "51f72883-b55f-4b66-a0e3-36d5cb3ceebc"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.829392 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "51f72883-b55f-4b66-a0e3-36d5cb3ceebc" (UID: "51f72883-b55f-4b66-a0e3-36d5cb3ceebc"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.830815 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-builder-dockercfg-q2gzr-push" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-push") pod "51f72883-b55f-4b66-a0e3-36d5cb3ceebc" (UID: "51f72883-b55f-4b66-a0e3-36d5cb3ceebc"). InnerVolumeSpecName "builder-dockercfg-q2gzr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.832039 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-kube-api-access-fbx9r" (OuterVolumeSpecName: "kube-api-access-fbx9r") pod "51f72883-b55f-4b66-a0e3-36d5cb3ceebc" (UID: "51f72883-b55f-4b66-a0e3-36d5cb3ceebc"). InnerVolumeSpecName "kube-api-access-fbx9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.834233 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-builder-dockercfg-q2gzr-pull" (OuterVolumeSpecName: "builder-dockercfg-q2gzr-pull") pod "51f72883-b55f-4b66-a0e3-36d5cb3ceebc" (UID: "51f72883-b55f-4b66-a0e3-36d5cb3ceebc"). InnerVolumeSpecName "builder-dockercfg-q2gzr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.922545 4983 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.922586 4983 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.922599 4983 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.922611 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbx9r\" (UniqueName: \"kubernetes.io/projected/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-kube-api-access-fbx9r\") on node \"crc\" DevicePath \"\"" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.922624 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.922636 4983 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.922647 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-pull\" (UniqueName: \"kubernetes.io/secret/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-builder-dockercfg-q2gzr-pull\") on node \"crc\" DevicePath \"\"" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.922661 4983 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.922675 4983 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-q2gzr-push\" (UniqueName: \"kubernetes.io/secret/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-builder-dockercfg-q2gzr-push\") on node \"crc\" DevicePath \"\"" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.922688 4983 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 18 00:32:01 crc kubenswrapper[4983]: I0318 00:32:01.947016 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "51f72883-b55f-4b66-a0e3-36d5cb3ceebc" (UID: "51f72883-b55f-4b66-a0e3-36d5cb3ceebc"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:32:02 crc kubenswrapper[4983]: I0318 00:32:02.023661 4983 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 18 00:32:02 crc kubenswrapper[4983]: I0318 00:32:02.280683 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563232-jl6dt" event={"ID":"5db34645-4368-4df2-9da6-f3b4b119a4cf","Type":"ContainerStarted","Data":"74f29dd94385cfe5b21027acdab148521bd9bc0f3ef6d55200dcc6e76181fea1"} Mar 18 00:32:02 crc kubenswrapper[4983]: I0318 00:32:02.283888 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"51f72883-b55f-4b66-a0e3-36d5cb3ceebc","Type":"ContainerDied","Data":"79b8c5f4ae1c38cecd1d6cf1205b62c85bbbf1ebe1d71717d8b873ffaeb710ae"} Mar 18 00:32:02 crc kubenswrapper[4983]: I0318 00:32:02.284012 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 18 00:32:02 crc kubenswrapper[4983]: I0318 00:32:02.284695 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79b8c5f4ae1c38cecd1d6cf1205b62c85bbbf1ebe1d71717d8b873ffaeb710ae" Mar 18 00:32:02 crc kubenswrapper[4983]: I0318 00:32:02.296586 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29563232-jl6dt" podStartSLOduration=1.3042277580000001 podStartE2EDuration="2.296563179s" podCreationTimestamp="2026-03-18 00:32:00 +0000 UTC" firstStartedPulling="2026-03-18 00:32:00.800252536 +0000 UTC m=+1438.697627166" lastFinishedPulling="2026-03-18 00:32:01.792587957 +0000 UTC m=+1439.689962587" observedRunningTime="2026-03-18 00:32:02.295837269 +0000 UTC m=+1440.193211869" watchObservedRunningTime="2026-03-18 00:32:02.296563179 +0000 UTC m=+1440.193937809" Mar 18 00:32:02 crc kubenswrapper[4983]: I0318 00:32:02.965732 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "51f72883-b55f-4b66-a0e3-36d5cb3ceebc" (UID: "51f72883-b55f-4b66-a0e3-36d5cb3ceebc"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:32:03 crc kubenswrapper[4983]: I0318 00:32:03.044172 4983 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/51f72883-b55f-4b66-a0e3-36d5cb3ceebc-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 18 00:32:03 crc kubenswrapper[4983]: I0318 00:32:03.296868 4983 generic.go:334] "Generic (PLEG): container finished" podID="5db34645-4368-4df2-9da6-f3b4b119a4cf" containerID="74f29dd94385cfe5b21027acdab148521bd9bc0f3ef6d55200dcc6e76181fea1" exitCode=0 Mar 18 00:32:03 crc kubenswrapper[4983]: I0318 00:32:03.297005 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563232-jl6dt" event={"ID":"5db34645-4368-4df2-9da6-f3b4b119a4cf","Type":"ContainerDied","Data":"74f29dd94385cfe5b21027acdab148521bd9bc0f3ef6d55200dcc6e76181fea1"} Mar 18 00:32:04 crc kubenswrapper[4983]: I0318 00:32:04.582473 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563232-jl6dt" Mar 18 00:32:04 crc kubenswrapper[4983]: I0318 00:32:04.770678 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w85cb\" (UniqueName: \"kubernetes.io/projected/5db34645-4368-4df2-9da6-f3b4b119a4cf-kube-api-access-w85cb\") pod \"5db34645-4368-4df2-9da6-f3b4b119a4cf\" (UID: \"5db34645-4368-4df2-9da6-f3b4b119a4cf\") " Mar 18 00:32:04 crc kubenswrapper[4983]: I0318 00:32:04.777116 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5db34645-4368-4df2-9da6-f3b4b119a4cf-kube-api-access-w85cb" (OuterVolumeSpecName: "kube-api-access-w85cb") pod "5db34645-4368-4df2-9da6-f3b4b119a4cf" (UID: "5db34645-4368-4df2-9da6-f3b4b119a4cf"). InnerVolumeSpecName "kube-api-access-w85cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:32:04 crc kubenswrapper[4983]: I0318 00:32:04.872856 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w85cb\" (UniqueName: \"kubernetes.io/projected/5db34645-4368-4df2-9da6-f3b4b119a4cf-kube-api-access-w85cb\") on node \"crc\" DevicePath \"\"" Mar 18 00:32:05 crc kubenswrapper[4983]: I0318 00:32:05.316122 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563232-jl6dt" event={"ID":"5db34645-4368-4df2-9da6-f3b4b119a4cf","Type":"ContainerDied","Data":"4e805c09efc5ff12da49751d8aa211ff59765f8dc251ac1966d668e21b753c9b"} Mar 18 00:32:05 crc kubenswrapper[4983]: I0318 00:32:05.316184 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e805c09efc5ff12da49751d8aa211ff59765f8dc251ac1966d668e21b753c9b" Mar 18 00:32:05 crc kubenswrapper[4983]: I0318 00:32:05.316270 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563232-jl6dt" Mar 18 00:32:05 crc kubenswrapper[4983]: I0318 00:32:05.402091 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563226-87lzm"] Mar 18 00:32:05 crc kubenswrapper[4983]: I0318 00:32:05.414185 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563226-87lzm"] Mar 18 00:32:06 crc kubenswrapper[4983]: I0318 00:32:06.854876 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="018dc6e9-49c7-443a-8f85-854247ef9d94" path="/var/lib/kubelet/pods/018dc6e9-49c7-443a-8f85-854247ef9d94/volumes" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.315407 4983 scope.go:117] "RemoveContainer" containerID="5553c9f8f36da1a7fa7e9e2d4fe6285578f7857a0609e4f1eb6ba3eccc6c0bc4" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.330439 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-79ff6f684b-49b74"] Mar 18 00:32:10 crc kubenswrapper[4983]: E0318 00:32:10.330747 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51f72883-b55f-4b66-a0e3-36d5cb3ceebc" containerName="git-clone" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.330763 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="51f72883-b55f-4b66-a0e3-36d5cb3ceebc" containerName="git-clone" Mar 18 00:32:10 crc kubenswrapper[4983]: E0318 00:32:10.330783 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51f72883-b55f-4b66-a0e3-36d5cb3ceebc" containerName="manage-dockerfile" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.330791 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="51f72883-b55f-4b66-a0e3-36d5cb3ceebc" containerName="manage-dockerfile" Mar 18 00:32:10 crc kubenswrapper[4983]: E0318 00:32:10.330807 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51f72883-b55f-4b66-a0e3-36d5cb3ceebc" containerName="docker-build" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.330815 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="51f72883-b55f-4b66-a0e3-36d5cb3ceebc" containerName="docker-build" Mar 18 00:32:10 crc kubenswrapper[4983]: E0318 00:32:10.330829 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5db34645-4368-4df2-9da6-f3b4b119a4cf" containerName="oc" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.330838 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5db34645-4368-4df2-9da6-f3b4b119a4cf" containerName="oc" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.331005 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="5db34645-4368-4df2-9da6-f3b4b119a4cf" containerName="oc" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.331016 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="51f72883-b55f-4b66-a0e3-36d5cb3ceebc" containerName="docker-build" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.333544 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-79ff6f684b-49b74" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.336159 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-operator-dockercfg-ffw75" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.360691 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-79ff6f684b-49b74"] Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.451753 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/71ccd954-e60f-4392-9ce6-3672ff3cadce-runner\") pod \"smart-gateway-operator-79ff6f684b-49b74\" (UID: \"71ccd954-e60f-4392-9ce6-3672ff3cadce\") " pod="service-telemetry/smart-gateway-operator-79ff6f684b-49b74" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.451864 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wmpn\" (UniqueName: \"kubernetes.io/projected/71ccd954-e60f-4392-9ce6-3672ff3cadce-kube-api-access-5wmpn\") pod \"smart-gateway-operator-79ff6f684b-49b74\" (UID: \"71ccd954-e60f-4392-9ce6-3672ff3cadce\") " pod="service-telemetry/smart-gateway-operator-79ff6f684b-49b74" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.552866 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/71ccd954-e60f-4392-9ce6-3672ff3cadce-runner\") pod \"smart-gateway-operator-79ff6f684b-49b74\" (UID: \"71ccd954-e60f-4392-9ce6-3672ff3cadce\") " pod="service-telemetry/smart-gateway-operator-79ff6f684b-49b74" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.552945 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wmpn\" (UniqueName: \"kubernetes.io/projected/71ccd954-e60f-4392-9ce6-3672ff3cadce-kube-api-access-5wmpn\") pod \"smart-gateway-operator-79ff6f684b-49b74\" (UID: \"71ccd954-e60f-4392-9ce6-3672ff3cadce\") " pod="service-telemetry/smart-gateway-operator-79ff6f684b-49b74" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.553792 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/71ccd954-e60f-4392-9ce6-3672ff3cadce-runner\") pod \"smart-gateway-operator-79ff6f684b-49b74\" (UID: \"71ccd954-e60f-4392-9ce6-3672ff3cadce\") " pod="service-telemetry/smart-gateway-operator-79ff6f684b-49b74" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.587178 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wmpn\" (UniqueName: \"kubernetes.io/projected/71ccd954-e60f-4392-9ce6-3672ff3cadce-kube-api-access-5wmpn\") pod \"smart-gateway-operator-79ff6f684b-49b74\" (UID: \"71ccd954-e60f-4392-9ce6-3672ff3cadce\") " pod="service-telemetry/smart-gateway-operator-79ff6f684b-49b74" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.678032 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-79ff6f684b-49b74" Mar 18 00:32:10 crc kubenswrapper[4983]: I0318 00:32:10.879818 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-79ff6f684b-49b74"] Mar 18 00:32:10 crc kubenswrapper[4983]: W0318 00:32:10.889279 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71ccd954_e60f_4392_9ce6_3672ff3cadce.slice/crio-60045280263eb1505f1ab07228edddc7cd391c0951d7101f9e3a5ca9f56335c1 WatchSource:0}: Error finding container 60045280263eb1505f1ab07228edddc7cd391c0951d7101f9e3a5ca9f56335c1: Status 404 returned error can't find the container with id 60045280263eb1505f1ab07228edddc7cd391c0951d7101f9e3a5ca9f56335c1 Mar 18 00:32:11 crc kubenswrapper[4983]: I0318 00:32:11.370799 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-79ff6f684b-49b74" event={"ID":"71ccd954-e60f-4392-9ce6-3672ff3cadce","Type":"ContainerStarted","Data":"60045280263eb1505f1ab07228edddc7cd391c0951d7101f9e3a5ca9f56335c1"} Mar 18 00:32:13 crc kubenswrapper[4983]: I0318 00:32:13.626391 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj"] Mar 18 00:32:13 crc kubenswrapper[4983]: I0318 00:32:13.627817 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj" Mar 18 00:32:13 crc kubenswrapper[4983]: I0318 00:32:13.629383 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-operator-dockercfg-m9g2z" Mar 18 00:32:13 crc kubenswrapper[4983]: I0318 00:32:13.645695 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj"] Mar 18 00:32:13 crc kubenswrapper[4983]: I0318 00:32:13.810319 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/6baa739d-8869-4888-95b8-821fa42715b7-runner\") pod \"service-telemetry-operator-7f455c87c8-7n4mj\" (UID: \"6baa739d-8869-4888-95b8-821fa42715b7\") " pod="service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj" Mar 18 00:32:13 crc kubenswrapper[4983]: I0318 00:32:13.810396 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhcrw\" (UniqueName: \"kubernetes.io/projected/6baa739d-8869-4888-95b8-821fa42715b7-kube-api-access-mhcrw\") pod \"service-telemetry-operator-7f455c87c8-7n4mj\" (UID: \"6baa739d-8869-4888-95b8-821fa42715b7\") " pod="service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj" Mar 18 00:32:13 crc kubenswrapper[4983]: I0318 00:32:13.911471 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/6baa739d-8869-4888-95b8-821fa42715b7-runner\") pod \"service-telemetry-operator-7f455c87c8-7n4mj\" (UID: \"6baa739d-8869-4888-95b8-821fa42715b7\") " pod="service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj" Mar 18 00:32:13 crc kubenswrapper[4983]: I0318 00:32:13.911548 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhcrw\" (UniqueName: \"kubernetes.io/projected/6baa739d-8869-4888-95b8-821fa42715b7-kube-api-access-mhcrw\") pod \"service-telemetry-operator-7f455c87c8-7n4mj\" (UID: \"6baa739d-8869-4888-95b8-821fa42715b7\") " pod="service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj" Mar 18 00:32:13 crc kubenswrapper[4983]: I0318 00:32:13.912198 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/6baa739d-8869-4888-95b8-821fa42715b7-runner\") pod \"service-telemetry-operator-7f455c87c8-7n4mj\" (UID: \"6baa739d-8869-4888-95b8-821fa42715b7\") " pod="service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj" Mar 18 00:32:13 crc kubenswrapper[4983]: I0318 00:32:13.957210 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhcrw\" (UniqueName: \"kubernetes.io/projected/6baa739d-8869-4888-95b8-821fa42715b7-kube-api-access-mhcrw\") pod \"service-telemetry-operator-7f455c87c8-7n4mj\" (UID: \"6baa739d-8869-4888-95b8-821fa42715b7\") " pod="service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj" Mar 18 00:32:14 crc kubenswrapper[4983]: I0318 00:32:14.246060 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj" Mar 18 00:32:22 crc kubenswrapper[4983]: I0318 00:32:22.496379 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj"] Mar 18 00:32:26 crc kubenswrapper[4983]: W0318 00:32:26.093567 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6baa739d_8869_4888_95b8_821fa42715b7.slice/crio-0cb48007b1052d7805f6c2463ebd68c5a837a75ba213d2eee2793b3231cc5a5a WatchSource:0}: Error finding container 0cb48007b1052d7805f6c2463ebd68c5a837a75ba213d2eee2793b3231cc5a5a: Status 404 returned error can't find the container with id 0cb48007b1052d7805f6c2463ebd68c5a837a75ba213d2eee2793b3231cc5a5a Mar 18 00:32:26 crc kubenswrapper[4983]: I0318 00:32:26.496303 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj" event={"ID":"6baa739d-8869-4888-95b8-821fa42715b7","Type":"ContainerStarted","Data":"0cb48007b1052d7805f6c2463ebd68c5a837a75ba213d2eee2793b3231cc5a5a"} Mar 18 00:32:26 crc kubenswrapper[4983]: E0318 00:32:26.646452 4983 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/smart-gateway-operator:stable-1.5" Mar 18 00:32:26 crc kubenswrapper[4983]: E0318 00:32:26.646643 4983 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/infrawatch/smart-gateway-operator:stable-1.5,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:smart-gateway-operator,ValueFrom:nil,},EnvVar{Name:ANSIBLE_GATHERING,Value:explicit,ValueFrom:nil,},EnvVar{Name:ANSIBLE_VERBOSITY_SMARTGATEWAY_SMARTGATEWAY_INFRA_WATCH,Value:4,ValueFrom:nil,},EnvVar{Name:ANSIBLE_DEBUG_LOGS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CORE_SMARTGATEWAY_IMAGE,Value:image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BRIDGE_SMARTGATEWAY_IMAGE,Value:image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-bridge:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OAUTH_PROXY_IMAGE,Value:quay.io/openshift/origin-oauth-proxy:latest,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:smart-gateway-operator.v5.0.1773793923,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:runner,ReadOnly:false,MountPath:/tmp/ansible-operator/runner,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5wmpn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod smart-gateway-operator-79ff6f684b-49b74_service-telemetry(71ccd954-e60f-4392-9ce6-3672ff3cadce): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 00:32:26 crc kubenswrapper[4983]: E0318 00:32:26.648241 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/smart-gateway-operator-79ff6f684b-49b74" podUID="71ccd954-e60f-4392-9ce6-3672ff3cadce" Mar 18 00:32:27 crc kubenswrapper[4983]: E0318 00:32:27.504920 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/smart-gateway-operator:stable-1.5\\\"\"" pod="service-telemetry/smart-gateway-operator-79ff6f684b-49b74" podUID="71ccd954-e60f-4392-9ce6-3672ff3cadce" Mar 18 00:32:32 crc kubenswrapper[4983]: I0318 00:32:32.547668 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj" event={"ID":"6baa739d-8869-4888-95b8-821fa42715b7","Type":"ContainerStarted","Data":"50fc7abb15a012580d844be765ce01b7eff4bce646f4372747089f08672849e9"} Mar 18 00:32:32 crc kubenswrapper[4983]: I0318 00:32:32.580227 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-7f455c87c8-7n4mj" podStartSLOduration=14.144393638 podStartE2EDuration="19.58020176s" podCreationTimestamp="2026-03-18 00:32:13 +0000 UTC" firstStartedPulling="2026-03-18 00:32:26.097197988 +0000 UTC m=+1463.994572638" lastFinishedPulling="2026-03-18 00:32:31.5330061 +0000 UTC m=+1469.430380760" observedRunningTime="2026-03-18 00:32:32.575381386 +0000 UTC m=+1470.472756036" watchObservedRunningTime="2026-03-18 00:32:32.58020176 +0000 UTC m=+1470.477576410" Mar 18 00:32:40 crc kubenswrapper[4983]: I0318 00:32:40.627419 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-79ff6f684b-49b74" event={"ID":"71ccd954-e60f-4392-9ce6-3672ff3cadce","Type":"ContainerStarted","Data":"cd726ba941c80ca5d82d0a3f80e09b234e9ff031989a9f4b7d3573a3bc510eb7"} Mar 18 00:32:40 crc kubenswrapper[4983]: I0318 00:32:40.661897 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-79ff6f684b-49b74" podStartSLOduration=1.959377828 podStartE2EDuration="30.66187639s" podCreationTimestamp="2026-03-18 00:32:10 +0000 UTC" firstStartedPulling="2026-03-18 00:32:10.892660593 +0000 UTC m=+1448.790035203" lastFinishedPulling="2026-03-18 00:32:39.595159145 +0000 UTC m=+1477.492533765" observedRunningTime="2026-03-18 00:32:40.656182001 +0000 UTC m=+1478.553556641" watchObservedRunningTime="2026-03-18 00:32:40.66187639 +0000 UTC m=+1478.559251010" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.749178 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-2rwvf"] Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.750847 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.759825 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.760149 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.760335 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.760502 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.760679 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.761345 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.761355 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-hgsq6" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.766857 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-2rwvf"] Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.904419 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/b26927e3-2404-4f34-9fd5-afb37a1f4387-sasl-config\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.904799 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwr7k\" (UniqueName: \"kubernetes.io/projected/b26927e3-2404-4f34-9fd5-afb37a1f4387-kube-api-access-cwr7k\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.904847 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.904887 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.904905 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-sasl-users\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.904982 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:51 crc kubenswrapper[4983]: I0318 00:32:51.905000 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.005770 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/b26927e3-2404-4f34-9fd5-afb37a1f4387-sasl-config\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.005860 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwr7k\" (UniqueName: \"kubernetes.io/projected/b26927e3-2404-4f34-9fd5-afb37a1f4387-kube-api-access-cwr7k\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.005902 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.005946 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.005969 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-sasl-users\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.006018 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.006036 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.006703 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/b26927e3-2404-4f34-9fd5-afb37a1f4387-sasl-config\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.011952 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.012414 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-sasl-users\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.012800 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.031072 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.031142 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwr7k\" (UniqueName: \"kubernetes.io/projected/b26927e3-2404-4f34-9fd5-afb37a1f4387-kube-api-access-cwr7k\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.035383 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-2rwvf\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.076860 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.554676 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-2rwvf"] Mar 18 00:32:52 crc kubenswrapper[4983]: W0318 00:32:52.566892 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb26927e3_2404_4f34_9fd5_afb37a1f4387.slice/crio-fd8d6f4b385a812b39f14b9a35923d5aebd703130084db58a0918a3f957c54ee WatchSource:0}: Error finding container fd8d6f4b385a812b39f14b9a35923d5aebd703130084db58a0918a3f957c54ee: Status 404 returned error can't find the container with id fd8d6f4b385a812b39f14b9a35923d5aebd703130084db58a0918a3f957c54ee Mar 18 00:32:52 crc kubenswrapper[4983]: I0318 00:32:52.733766 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" event={"ID":"b26927e3-2404-4f34-9fd5-afb37a1f4387","Type":"ContainerStarted","Data":"fd8d6f4b385a812b39f14b9a35923d5aebd703130084db58a0918a3f957c54ee"} Mar 18 00:32:59 crc kubenswrapper[4983]: I0318 00:32:59.795596 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" event={"ID":"b26927e3-2404-4f34-9fd5-afb37a1f4387","Type":"ContainerStarted","Data":"47a42d8b569bee2ab9848da5a9e7e5a310cd2380de8e205a4eeccb5220929297"} Mar 18 00:32:59 crc kubenswrapper[4983]: I0318 00:32:59.839079 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" podStartSLOduration=2.303488568 podStartE2EDuration="8.839057227s" podCreationTimestamp="2026-03-18 00:32:51 +0000 UTC" firstStartedPulling="2026-03-18 00:32:52.570177631 +0000 UTC m=+1490.467552251" lastFinishedPulling="2026-03-18 00:32:59.1057463 +0000 UTC m=+1497.003120910" observedRunningTime="2026-03-18 00:32:59.825832128 +0000 UTC m=+1497.723206818" watchObservedRunningTime="2026-03-18 00:32:59.839057227 +0000 UTC m=+1497.736431857" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.172043 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.174251 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.176892 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.176904 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-stf-dockercfg-9k2nc" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.177742 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-2" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.178906 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-1" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.179160 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-prometheus-proxy-tls" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.179486 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.179700 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"serving-certs-ca-bundle" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.179902 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-session-secret" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.180041 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-tls-assets-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.189082 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-web-config" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.204917 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.369965 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec767daa-3f6a-4501-b63a-dd8949ef6199-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.370060 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.370105 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0925492b-49f0-4478-8f07-837ff5e027a3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0925492b-49f0-4478-8f07-837ff5e027a3\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.370273 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ec767daa-3f6a-4501-b63a-dd8949ef6199-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.370331 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/ec767daa-3f6a-4501-b63a-dd8949ef6199-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.370411 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.370446 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/ec767daa-3f6a-4501-b63a-dd8949ef6199-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.370477 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ec767daa-3f6a-4501-b63a-dd8949ef6199-tls-assets\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.370508 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-config\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.370653 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ec767daa-3f6a-4501-b63a-dd8949ef6199-config-out\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.370711 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-web-config\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.370755 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrb4t\" (UniqueName: \"kubernetes.io/projected/ec767daa-3f6a-4501-b63a-dd8949ef6199-kube-api-access-mrb4t\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.472509 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec767daa-3f6a-4501-b63a-dd8949ef6199-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.472938 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.472979 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0925492b-49f0-4478-8f07-837ff5e027a3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0925492b-49f0-4478-8f07-837ff5e027a3\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.473023 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ec767daa-3f6a-4501-b63a-dd8949ef6199-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.473049 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/ec767daa-3f6a-4501-b63a-dd8949ef6199-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.473084 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.473240 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/ec767daa-3f6a-4501-b63a-dd8949ef6199-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: E0318 00:33:02.473366 4983 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Mar 18 00:33:02 crc kubenswrapper[4983]: E0318 00:33:02.473478 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-secret-default-prometheus-proxy-tls podName:ec767daa-3f6a-4501-b63a-dd8949ef6199 nodeName:}" failed. No retries permitted until 2026-03-18 00:33:02.973447165 +0000 UTC m=+1500.870821815 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "ec767daa-3f6a-4501-b63a-dd8949ef6199") : secret "default-prometheus-proxy-tls" not found Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.473890 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ec767daa-3f6a-4501-b63a-dd8949ef6199-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.474174 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/ec767daa-3f6a-4501-b63a-dd8949ef6199-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.474363 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ec767daa-3f6a-4501-b63a-dd8949ef6199-tls-assets\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.474414 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-config\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.474456 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ec767daa-3f6a-4501-b63a-dd8949ef6199-config-out\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.474480 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/ec767daa-3f6a-4501-b63a-dd8949ef6199-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.474503 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-web-config\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.474622 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrb4t\" (UniqueName: \"kubernetes.io/projected/ec767daa-3f6a-4501-b63a-dd8949ef6199-kube-api-access-mrb4t\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.474976 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec767daa-3f6a-4501-b63a-dd8949ef6199-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.480452 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-web-config\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.480648 4983 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.480672 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0925492b-49f0-4478-8f07-837ff5e027a3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0925492b-49f0-4478-8f07-837ff5e027a3\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0eeef6d1f39b64e736ffefad917219d3c7b5fd0fc5d419d884a81044c75ac137/globalmount\"" pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.480729 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-config\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.481597 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ec767daa-3f6a-4501-b63a-dd8949ef6199-config-out\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.491556 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.501042 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ec767daa-3f6a-4501-b63a-dd8949ef6199-tls-assets\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.511228 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrb4t\" (UniqueName: \"kubernetes.io/projected/ec767daa-3f6a-4501-b63a-dd8949ef6199-kube-api-access-mrb4t\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.514692 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0925492b-49f0-4478-8f07-837ff5e027a3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0925492b-49f0-4478-8f07-837ff5e027a3\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: I0318 00:33:02.983176 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:02 crc kubenswrapper[4983]: E0318 00:33:02.983715 4983 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Mar 18 00:33:02 crc kubenswrapper[4983]: E0318 00:33:02.983796 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-secret-default-prometheus-proxy-tls podName:ec767daa-3f6a-4501-b63a-dd8949ef6199 nodeName:}" failed. No retries permitted until 2026-03-18 00:33:03.983773844 +0000 UTC m=+1501.881148484 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "ec767daa-3f6a-4501-b63a-dd8949ef6199") : secret "default-prometheus-proxy-tls" not found Mar 18 00:33:03 crc kubenswrapper[4983]: I0318 00:33:03.997065 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:04 crc kubenswrapper[4983]: I0318 00:33:04.008119 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ec767daa-3f6a-4501-b63a-dd8949ef6199-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"ec767daa-3f6a-4501-b63a-dd8949ef6199\") " pod="service-telemetry/prometheus-default-0" Mar 18 00:33:04 crc kubenswrapper[4983]: I0318 00:33:04.303204 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-stf-dockercfg-9k2nc" Mar 18 00:33:04 crc kubenswrapper[4983]: I0318 00:33:04.312423 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Mar 18 00:33:04 crc kubenswrapper[4983]: W0318 00:33:04.783334 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec767daa_3f6a_4501_b63a_dd8949ef6199.slice/crio-867e696fcfbf5f94bc8ae7408694dc169bf12a0c0f9a41f318a3b32c0b4de87d WatchSource:0}: Error finding container 867e696fcfbf5f94bc8ae7408694dc169bf12a0c0f9a41f318a3b32c0b4de87d: Status 404 returned error can't find the container with id 867e696fcfbf5f94bc8ae7408694dc169bf12a0c0f9a41f318a3b32c0b4de87d Mar 18 00:33:04 crc kubenswrapper[4983]: I0318 00:33:04.800143 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 18 00:33:04 crc kubenswrapper[4983]: I0318 00:33:04.842742 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"ec767daa-3f6a-4501-b63a-dd8949ef6199","Type":"ContainerStarted","Data":"867e696fcfbf5f94bc8ae7408694dc169bf12a0c0f9a41f318a3b32c0b4de87d"} Mar 18 00:33:09 crc kubenswrapper[4983]: I0318 00:33:09.894507 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"ec767daa-3f6a-4501-b63a-dd8949ef6199","Type":"ContainerStarted","Data":"c926e01e949bc3ceaf97dab3274e6dfc593f9cc31f75098e4d76c75d7160522c"} Mar 18 00:33:12 crc kubenswrapper[4983]: I0318 00:33:12.608279 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-mqksz"] Mar 18 00:33:12 crc kubenswrapper[4983]: I0318 00:33:12.609461 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6856cfb745-mqksz" Mar 18 00:33:12 crc kubenswrapper[4983]: I0318 00:33:12.626369 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-mqksz"] Mar 18 00:33:12 crc kubenswrapper[4983]: I0318 00:33:12.764815 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l46cr\" (UniqueName: \"kubernetes.io/projected/c025a69f-f942-4648-bb53-972903a8e14d-kube-api-access-l46cr\") pod \"default-snmp-webhook-6856cfb745-mqksz\" (UID: \"c025a69f-f942-4648-bb53-972903a8e14d\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-mqksz" Mar 18 00:33:12 crc kubenswrapper[4983]: I0318 00:33:12.866661 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l46cr\" (UniqueName: \"kubernetes.io/projected/c025a69f-f942-4648-bb53-972903a8e14d-kube-api-access-l46cr\") pod \"default-snmp-webhook-6856cfb745-mqksz\" (UID: \"c025a69f-f942-4648-bb53-972903a8e14d\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-mqksz" Mar 18 00:33:12 crc kubenswrapper[4983]: I0318 00:33:12.888035 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l46cr\" (UniqueName: \"kubernetes.io/projected/c025a69f-f942-4648-bb53-972903a8e14d-kube-api-access-l46cr\") pod \"default-snmp-webhook-6856cfb745-mqksz\" (UID: \"c025a69f-f942-4648-bb53-972903a8e14d\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-mqksz" Mar 18 00:33:12 crc kubenswrapper[4983]: I0318 00:33:12.942503 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6856cfb745-mqksz" Mar 18 00:33:13 crc kubenswrapper[4983]: I0318 00:33:13.249309 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-mqksz"] Mar 18 00:33:13 crc kubenswrapper[4983]: I0318 00:33:13.258945 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:33:13 crc kubenswrapper[4983]: I0318 00:33:13.258978 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:33:13 crc kubenswrapper[4983]: I0318 00:33:13.939855 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6856cfb745-mqksz" event={"ID":"c025a69f-f942-4648-bb53-972903a8e14d","Type":"ContainerStarted","Data":"5cef9b1cbc0ba1b3a2d05dea82c2ca6b9f9d368f539de884545848ad7bd1a3ac"} Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.038075 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.040649 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.043443 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-stf-dockercfg-ccx4g" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.043765 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-alertmanager-proxy-tls" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.043905 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.044017 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-tls-assets-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.044099 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-web-config" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.044196 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-cluster-tls-config" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.046585 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-generated" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.238977 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-config-volume\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.239022 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.239176 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-006bb3b2-4f37-4a67-8c1a-a05d6eba04b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-006bb3b2-4f37-4a67-8c1a-a05d6eba04b9\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.239253 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.239379 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6c1ecce7-261c-4f82-863c-641ef76200ae-tls-assets\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.239425 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6c1ecce7-261c-4f82-863c-641ef76200ae-config-out\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.239483 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.239560 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-web-config\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.239671 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsl22\" (UniqueName: \"kubernetes.io/projected/6c1ecce7-261c-4f82-863c-641ef76200ae-kube-api-access-jsl22\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.341021 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.341074 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6c1ecce7-261c-4f82-863c-641ef76200ae-tls-assets\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.341110 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6c1ecce7-261c-4f82-863c-641ef76200ae-config-out\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.341136 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.341163 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-web-config\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.341199 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsl22\" (UniqueName: \"kubernetes.io/projected/6c1ecce7-261c-4f82-863c-641ef76200ae-kube-api-access-jsl22\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.341230 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-config-volume\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.341250 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.341293 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-006bb3b2-4f37-4a67-8c1a-a05d6eba04b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-006bb3b2-4f37-4a67-8c1a-a05d6eba04b9\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: E0318 00:33:16.341512 4983 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Mar 18 00:33:16 crc kubenswrapper[4983]: E0318 00:33:16.341591 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-alertmanager-proxy-tls podName:6c1ecce7-261c-4f82-863c-641ef76200ae nodeName:}" failed. No retries permitted until 2026-03-18 00:33:16.841572715 +0000 UTC m=+1514.738947325 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "6c1ecce7-261c-4f82-863c-641ef76200ae") : secret "default-alertmanager-proxy-tls" not found Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.348298 4983 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.348596 4983 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-006bb3b2-4f37-4a67-8c1a-a05d6eba04b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-006bb3b2-4f37-4a67-8c1a-a05d6eba04b9\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b47373eaeef5d82ec30ce70b12c553dca074c3dbe3b16db54dee5003d87124db/globalmount\"" pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.352071 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6c1ecce7-261c-4f82-863c-641ef76200ae-config-out\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.352176 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6c1ecce7-261c-4f82-863c-641ef76200ae-tls-assets\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.352281 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-web-config\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.357675 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.358702 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-config-volume\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.361377 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.362153 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsl22\" (UniqueName: \"kubernetes.io/projected/6c1ecce7-261c-4f82-863c-641ef76200ae-kube-api-access-jsl22\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.409115 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-006bb3b2-4f37-4a67-8c1a-a05d6eba04b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-006bb3b2-4f37-4a67-8c1a-a05d6eba04b9\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.847330 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:16 crc kubenswrapper[4983]: E0318 00:33:16.847534 4983 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Mar 18 00:33:16 crc kubenswrapper[4983]: E0318 00:33:16.848061 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-alertmanager-proxy-tls podName:6c1ecce7-261c-4f82-863c-641ef76200ae nodeName:}" failed. No retries permitted until 2026-03-18 00:33:17.848033497 +0000 UTC m=+1515.745408137 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "6c1ecce7-261c-4f82-863c-641ef76200ae") : secret "default-alertmanager-proxy-tls" not found Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.964705 4983 generic.go:334] "Generic (PLEG): container finished" podID="ec767daa-3f6a-4501-b63a-dd8949ef6199" containerID="c926e01e949bc3ceaf97dab3274e6dfc593f9cc31f75098e4d76c75d7160522c" exitCode=0 Mar 18 00:33:16 crc kubenswrapper[4983]: I0318 00:33:16.964753 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"ec767daa-3f6a-4501-b63a-dd8949ef6199","Type":"ContainerDied","Data":"c926e01e949bc3ceaf97dab3274e6dfc593f9cc31f75098e4d76c75d7160522c"} Mar 18 00:33:17 crc kubenswrapper[4983]: I0318 00:33:17.861242 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:17 crc kubenswrapper[4983]: E0318 00:33:17.861459 4983 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Mar 18 00:33:17 crc kubenswrapper[4983]: E0318 00:33:17.861547 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-alertmanager-proxy-tls podName:6c1ecce7-261c-4f82-863c-641ef76200ae nodeName:}" failed. No retries permitted until 2026-03-18 00:33:19.861525676 +0000 UTC m=+1517.758900286 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "6c1ecce7-261c-4f82-863c-641ef76200ae") : secret "default-alertmanager-proxy-tls" not found Mar 18 00:33:19 crc kubenswrapper[4983]: I0318 00:33:19.891875 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:19 crc kubenswrapper[4983]: I0318 00:33:19.897705 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c1ecce7-261c-4f82-863c-641ef76200ae-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"6c1ecce7-261c-4f82-863c-641ef76200ae\") " pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:19 crc kubenswrapper[4983]: I0318 00:33:19.959162 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Mar 18 00:33:20 crc kubenswrapper[4983]: I0318 00:33:20.963428 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 18 00:33:20 crc kubenswrapper[4983]: I0318 00:33:20.997280 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6856cfb745-mqksz" event={"ID":"c025a69f-f942-4648-bb53-972903a8e14d","Type":"ContainerStarted","Data":"91defddff066d544e1158d2cd63ac438e48c260112a8c7a5072981c097d24d66"} Mar 18 00:33:21 crc kubenswrapper[4983]: I0318 00:33:21.015451 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-snmp-webhook-6856cfb745-mqksz" podStartSLOduration=1.6896160569999998 podStartE2EDuration="9.01543252s" podCreationTimestamp="2026-03-18 00:33:12 +0000 UTC" firstStartedPulling="2026-03-18 00:33:13.25827609 +0000 UTC m=+1511.155650700" lastFinishedPulling="2026-03-18 00:33:20.584092553 +0000 UTC m=+1518.481467163" observedRunningTime="2026-03-18 00:33:21.013198318 +0000 UTC m=+1518.910572938" watchObservedRunningTime="2026-03-18 00:33:21.01543252 +0000 UTC m=+1518.912807130" Mar 18 00:33:21 crc kubenswrapper[4983]: W0318 00:33:21.080236 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c1ecce7_261c_4f82_863c_641ef76200ae.slice/crio-1f0b33c6938cd41e8a8cd07c546e6b8ffe47afef93d209368375ec2d9a8d4e7b WatchSource:0}: Error finding container 1f0b33c6938cd41e8a8cd07c546e6b8ffe47afef93d209368375ec2d9a8d4e7b: Status 404 returned error can't find the container with id 1f0b33c6938cd41e8a8cd07c546e6b8ffe47afef93d209368375ec2d9a8d4e7b Mar 18 00:33:22 crc kubenswrapper[4983]: I0318 00:33:22.016072 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"6c1ecce7-261c-4f82-863c-641ef76200ae","Type":"ContainerStarted","Data":"1f0b33c6938cd41e8a8cd07c546e6b8ffe47afef93d209368375ec2d9a8d4e7b"} Mar 18 00:33:24 crc kubenswrapper[4983]: I0318 00:33:24.031207 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"6c1ecce7-261c-4f82-863c-641ef76200ae","Type":"ContainerStarted","Data":"961d0e3e42ff71b8e3300559e7b2bf5111560469a713faea84a0b8e1fbbfda11"} Mar 18 00:33:25 crc kubenswrapper[4983]: I0318 00:33:25.041957 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"ec767daa-3f6a-4501-b63a-dd8949ef6199","Type":"ContainerStarted","Data":"9c5c8811f2c5a8ebfc2c0c6421a88486da19adb9c6cb8c3765e020aad0e02be3"} Mar 18 00:33:27 crc kubenswrapper[4983]: I0318 00:33:27.059597 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"ec767daa-3f6a-4501-b63a-dd8949ef6199","Type":"ContainerStarted","Data":"32dfabd1bc287a9764ea73f0f2670244c3f272595b8be84f7777b29a997d9360"} Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.562438 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq"] Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.564903 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.568843 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-coll-meter-proxy-tls" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.574361 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-dockercfg-nclbj" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.574580 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-session-secret" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.574711 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-meter-sg-core-configmap" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.577327 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq"] Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.628856 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fszx9\" (UniqueName: \"kubernetes.io/projected/65a13175-2c14-45b7-92f5-4677c93be37f-kube-api-access-fszx9\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.628951 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/65a13175-2c14-45b7-92f5-4677c93be37f-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.628981 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/65a13175-2c14-45b7-92f5-4677c93be37f-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.629022 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/65a13175-2c14-45b7-92f5-4677c93be37f-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.629051 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/65a13175-2c14-45b7-92f5-4677c93be37f-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.730084 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fszx9\" (UniqueName: \"kubernetes.io/projected/65a13175-2c14-45b7-92f5-4677c93be37f-kube-api-access-fszx9\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.730154 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/65a13175-2c14-45b7-92f5-4677c93be37f-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.730178 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/65a13175-2c14-45b7-92f5-4677c93be37f-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.730222 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/65a13175-2c14-45b7-92f5-4677c93be37f-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.730254 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/65a13175-2c14-45b7-92f5-4677c93be37f-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.731225 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/65a13175-2c14-45b7-92f5-4677c93be37f-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: E0318 00:33:29.731571 4983 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Mar 18 00:33:29 crc kubenswrapper[4983]: E0318 00:33:29.731615 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/65a13175-2c14-45b7-92f5-4677c93be37f-default-cloud1-coll-meter-proxy-tls podName:65a13175-2c14-45b7-92f5-4677c93be37f nodeName:}" failed. No retries permitted until 2026-03-18 00:33:30.231602883 +0000 UTC m=+1528.128977493 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/65a13175-2c14-45b7-92f5-4677c93be37f-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" (UID: "65a13175-2c14-45b7-92f5-4677c93be37f") : secret "default-cloud1-coll-meter-proxy-tls" not found Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.731886 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/65a13175-2c14-45b7-92f5-4677c93be37f-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.737472 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/65a13175-2c14-45b7-92f5-4677c93be37f-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:29 crc kubenswrapper[4983]: I0318 00:33:29.749773 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fszx9\" (UniqueName: \"kubernetes.io/projected/65a13175-2c14-45b7-92f5-4677c93be37f-kube-api-access-fszx9\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:30 crc kubenswrapper[4983]: I0318 00:33:30.084228 4983 generic.go:334] "Generic (PLEG): container finished" podID="6c1ecce7-261c-4f82-863c-641ef76200ae" containerID="961d0e3e42ff71b8e3300559e7b2bf5111560469a713faea84a0b8e1fbbfda11" exitCode=0 Mar 18 00:33:30 crc kubenswrapper[4983]: I0318 00:33:30.084314 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"6c1ecce7-261c-4f82-863c-641ef76200ae","Type":"ContainerDied","Data":"961d0e3e42ff71b8e3300559e7b2bf5111560469a713faea84a0b8e1fbbfda11"} Mar 18 00:33:30 crc kubenswrapper[4983]: I0318 00:33:30.239061 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/65a13175-2c14-45b7-92f5-4677c93be37f-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:30 crc kubenswrapper[4983]: E0318 00:33:30.239207 4983 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Mar 18 00:33:30 crc kubenswrapper[4983]: E0318 00:33:30.239305 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/65a13175-2c14-45b7-92f5-4677c93be37f-default-cloud1-coll-meter-proxy-tls podName:65a13175-2c14-45b7-92f5-4677c93be37f nodeName:}" failed. No retries permitted until 2026-03-18 00:33:31.23928502 +0000 UTC m=+1529.136659630 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/65a13175-2c14-45b7-92f5-4677c93be37f-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" (UID: "65a13175-2c14-45b7-92f5-4677c93be37f") : secret "default-cloud1-coll-meter-proxy-tls" not found Mar 18 00:33:31 crc kubenswrapper[4983]: I0318 00:33:31.252244 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/65a13175-2c14-45b7-92f5-4677c93be37f-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:31 crc kubenswrapper[4983]: I0318 00:33:31.258743 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/65a13175-2c14-45b7-92f5-4677c93be37f-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq\" (UID: \"65a13175-2c14-45b7-92f5-4677c93be37f\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:31 crc kubenswrapper[4983]: I0318 00:33:31.401913 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.468601 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7"] Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.470985 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.473227 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-ceil-meter-proxy-tls" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.473612 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-meter-sg-core-configmap" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.478261 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7"] Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.571962 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/937b4985-05d5-48c0-b408-3164619e74ca-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.572024 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t65v6\" (UniqueName: \"kubernetes.io/projected/937b4985-05d5-48c0-b408-3164619e74ca-kube-api-access-t65v6\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.572071 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/937b4985-05d5-48c0-b408-3164619e74ca-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.572351 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/937b4985-05d5-48c0-b408-3164619e74ca-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.572388 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/937b4985-05d5-48c0-b408-3164619e74ca-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.674001 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/937b4985-05d5-48c0-b408-3164619e74ca-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.674041 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/937b4985-05d5-48c0-b408-3164619e74ca-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.674099 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/937b4985-05d5-48c0-b408-3164619e74ca-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.674131 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t65v6\" (UniqueName: \"kubernetes.io/projected/937b4985-05d5-48c0-b408-3164619e74ca-kube-api-access-t65v6\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.674171 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/937b4985-05d5-48c0-b408-3164619e74ca-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: E0318 00:33:32.674320 4983 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 18 00:33:32 crc kubenswrapper[4983]: E0318 00:33:32.674449 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/937b4985-05d5-48c0-b408-3164619e74ca-default-cloud1-ceil-meter-proxy-tls podName:937b4985-05d5-48c0-b408-3164619e74ca nodeName:}" failed. No retries permitted until 2026-03-18 00:33:33.174420591 +0000 UTC m=+1531.071795241 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/937b4985-05d5-48c0-b408-3164619e74ca-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" (UID: "937b4985-05d5-48c0-b408-3164619e74ca") : secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.674536 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/937b4985-05d5-48c0-b408-3164619e74ca-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.675129 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/937b4985-05d5-48c0-b408-3164619e74ca-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.683539 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/937b4985-05d5-48c0-b408-3164619e74ca-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:32 crc kubenswrapper[4983]: I0318 00:33:32.691130 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t65v6\" (UniqueName: \"kubernetes.io/projected/937b4985-05d5-48c0-b408-3164619e74ca-kube-api-access-t65v6\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:33 crc kubenswrapper[4983]: I0318 00:33:33.181455 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/937b4985-05d5-48c0-b408-3164619e74ca-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:33 crc kubenswrapper[4983]: E0318 00:33:33.181647 4983 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 18 00:33:33 crc kubenswrapper[4983]: E0318 00:33:33.181732 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/937b4985-05d5-48c0-b408-3164619e74ca-default-cloud1-ceil-meter-proxy-tls podName:937b4985-05d5-48c0-b408-3164619e74ca nodeName:}" failed. No retries permitted until 2026-03-18 00:33:34.181712576 +0000 UTC m=+1532.079087186 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/937b4985-05d5-48c0-b408-3164619e74ca-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" (UID: "937b4985-05d5-48c0-b408-3164619e74ca") : secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 18 00:33:34 crc kubenswrapper[4983]: I0318 00:33:34.195817 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/937b4985-05d5-48c0-b408-3164619e74ca-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:34 crc kubenswrapper[4983]: I0318 00:33:34.219955 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/937b4985-05d5-48c0-b408-3164619e74ca-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7\" (UID: \"937b4985-05d5-48c0-b408-3164619e74ca\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:34 crc kubenswrapper[4983]: I0318 00:33:34.289152 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.660609 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp"] Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.662865 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.667376 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-sens-meter-proxy-tls" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.667396 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-sens-meter-sg-core-configmap" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.672806 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp"] Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.717355 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/3753b92e-109f-4875-a1fe-70ea96a52a29-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.717398 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/3753b92e-109f-4875-a1fe-70ea96a52a29-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.717436 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/3753b92e-109f-4875-a1fe-70ea96a52a29-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.717453 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6cgd\" (UniqueName: \"kubernetes.io/projected/3753b92e-109f-4875-a1fe-70ea96a52a29-kube-api-access-p6cgd\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.717477 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/3753b92e-109f-4875-a1fe-70ea96a52a29-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.818698 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/3753b92e-109f-4875-a1fe-70ea96a52a29-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.819103 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/3753b92e-109f-4875-a1fe-70ea96a52a29-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: E0318 00:33:35.818888 4983 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.819186 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/3753b92e-109f-4875-a1fe-70ea96a52a29-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: E0318 00:33:35.819198 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3753b92e-109f-4875-a1fe-70ea96a52a29-default-cloud1-sens-meter-proxy-tls podName:3753b92e-109f-4875-a1fe-70ea96a52a29 nodeName:}" failed. No retries permitted until 2026-03-18 00:33:36.319177649 +0000 UTC m=+1534.216552259 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/3753b92e-109f-4875-a1fe-70ea96a52a29-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" (UID: "3753b92e-109f-4875-a1fe-70ea96a52a29") : secret "default-cloud1-sens-meter-proxy-tls" not found Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.819269 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6cgd\" (UniqueName: \"kubernetes.io/projected/3753b92e-109f-4875-a1fe-70ea96a52a29-kube-api-access-p6cgd\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.819886 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/3753b92e-109f-4875-a1fe-70ea96a52a29-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.820024 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/3753b92e-109f-4875-a1fe-70ea96a52a29-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.820481 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/3753b92e-109f-4875-a1fe-70ea96a52a29-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.830611 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/3753b92e-109f-4875-a1fe-70ea96a52a29-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:35 crc kubenswrapper[4983]: I0318 00:33:35.841924 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6cgd\" (UniqueName: \"kubernetes.io/projected/3753b92e-109f-4875-a1fe-70ea96a52a29-kube-api-access-p6cgd\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:36 crc kubenswrapper[4983]: I0318 00:33:36.225291 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq"] Mar 18 00:33:36 crc kubenswrapper[4983]: I0318 00:33:36.326752 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/3753b92e-109f-4875-a1fe-70ea96a52a29-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:36 crc kubenswrapper[4983]: E0318 00:33:36.326946 4983 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Mar 18 00:33:36 crc kubenswrapper[4983]: E0318 00:33:36.327021 4983 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3753b92e-109f-4875-a1fe-70ea96a52a29-default-cloud1-sens-meter-proxy-tls podName:3753b92e-109f-4875-a1fe-70ea96a52a29 nodeName:}" failed. No retries permitted until 2026-03-18 00:33:37.32700361 +0000 UTC m=+1535.224378220 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/3753b92e-109f-4875-a1fe-70ea96a52a29-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" (UID: "3753b92e-109f-4875-a1fe-70ea96a52a29") : secret "default-cloud1-sens-meter-proxy-tls" not found Mar 18 00:33:37 crc kubenswrapper[4983]: I0318 00:33:37.137970 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" event={"ID":"65a13175-2c14-45b7-92f5-4677c93be37f","Type":"ContainerStarted","Data":"150ab1f04be2e3d6d835fe745b18c1950536e451f5f200bb1fc4919da25aee93"} Mar 18 00:33:37 crc kubenswrapper[4983]: I0318 00:33:37.209775 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7"] Mar 18 00:33:37 crc kubenswrapper[4983]: I0318 00:33:37.338830 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/3753b92e-109f-4875-a1fe-70ea96a52a29-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:37 crc kubenswrapper[4983]: I0318 00:33:37.353706 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/3753b92e-109f-4875-a1fe-70ea96a52a29-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp\" (UID: \"3753b92e-109f-4875-a1fe-70ea96a52a29\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:37 crc kubenswrapper[4983]: I0318 00:33:37.529399 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" Mar 18 00:33:37 crc kubenswrapper[4983]: W0318 00:33:37.628883 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod937b4985_05d5_48c0_b408_3164619e74ca.slice/crio-76206baf22a0f4ad6763aaeffd50f0ceab7f4b8b6387eae910499ddf2f3dcdf3 WatchSource:0}: Error finding container 76206baf22a0f4ad6763aaeffd50f0ceab7f4b8b6387eae910499ddf2f3dcdf3: Status 404 returned error can't find the container with id 76206baf22a0f4ad6763aaeffd50f0ceab7f4b8b6387eae910499ddf2f3dcdf3 Mar 18 00:33:38 crc kubenswrapper[4983]: I0318 00:33:38.147675 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" event={"ID":"937b4985-05d5-48c0-b408-3164619e74ca","Type":"ContainerStarted","Data":"76206baf22a0f4ad6763aaeffd50f0ceab7f4b8b6387eae910499ddf2f3dcdf3"} Mar 18 00:33:38 crc kubenswrapper[4983]: I0318 00:33:38.479145 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp"] Mar 18 00:33:38 crc kubenswrapper[4983]: W0318 00:33:38.508772 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3753b92e_109f_4875_a1fe_70ea96a52a29.slice/crio-79514e6645592449751ad16d0684479f036d49560603fdaf39bbb848604a839e WatchSource:0}: Error finding container 79514e6645592449751ad16d0684479f036d49560603fdaf39bbb848604a839e: Status 404 returned error can't find the container with id 79514e6645592449751ad16d0684479f036d49560603fdaf39bbb848604a839e Mar 18 00:33:39 crc kubenswrapper[4983]: I0318 00:33:39.155668 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" event={"ID":"3753b92e-109f-4875-a1fe-70ea96a52a29","Type":"ContainerStarted","Data":"79514e6645592449751ad16d0684479f036d49560603fdaf39bbb848604a839e"} Mar 18 00:33:39 crc kubenswrapper[4983]: I0318 00:33:39.158406 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"6c1ecce7-261c-4f82-863c-641ef76200ae","Type":"ContainerStarted","Data":"9695ac2a2179e6b868e50eec4e9be3acf90fac800224a762820d8c0f3b003900"} Mar 18 00:33:39 crc kubenswrapper[4983]: I0318 00:33:39.160100 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"ec767daa-3f6a-4501-b63a-dd8949ef6199","Type":"ContainerStarted","Data":"8c76a5ca6d32535966021f18f461d03de507d396c3cf36db730604457a43b60e"} Mar 18 00:33:39 crc kubenswrapper[4983]: I0318 00:33:39.163755 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" event={"ID":"65a13175-2c14-45b7-92f5-4677c93be37f","Type":"ContainerStarted","Data":"412803cdbc034178a28c5a027668d1239332a62d15a459b7802f490c81c9151c"} Mar 18 00:33:39 crc kubenswrapper[4983]: I0318 00:33:39.181692 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-default-0" podStartSLOduration=4.858359899 podStartE2EDuration="38.181678089s" podCreationTimestamp="2026-03-18 00:33:01 +0000 UTC" firstStartedPulling="2026-03-18 00:33:04.78571013 +0000 UTC m=+1502.683084780" lastFinishedPulling="2026-03-18 00:33:38.10902836 +0000 UTC m=+1536.006402970" observedRunningTime="2026-03-18 00:33:39.179288113 +0000 UTC m=+1537.076662723" watchObservedRunningTime="2026-03-18 00:33:39.181678089 +0000 UTC m=+1537.079052699" Mar 18 00:33:39 crc kubenswrapper[4983]: I0318 00:33:39.313019 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/prometheus-default-0" Mar 18 00:33:40 crc kubenswrapper[4983]: I0318 00:33:40.172253 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" event={"ID":"937b4985-05d5-48c0-b408-3164619e74ca","Type":"ContainerStarted","Data":"ab45076cc85114c7c2402083464275316cc1881b0644bf60a56a6132ec1a1ad0"} Mar 18 00:33:40 crc kubenswrapper[4983]: I0318 00:33:40.172494 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" event={"ID":"937b4985-05d5-48c0-b408-3164619e74ca","Type":"ContainerStarted","Data":"5daad4c3c743e9226c3e2157d5e06c22f85aa615705028b0e3b65c1e573e6fbf"} Mar 18 00:33:40 crc kubenswrapper[4983]: I0318 00:33:40.174947 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" event={"ID":"3753b92e-109f-4875-a1fe-70ea96a52a29","Type":"ContainerStarted","Data":"fa76acf257a280301a6ed373efe8bb5a415d8b676f1b3f1accbe51f8f50f936c"} Mar 18 00:33:41 crc kubenswrapper[4983]: I0318 00:33:41.191481 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" event={"ID":"65a13175-2c14-45b7-92f5-4677c93be37f","Type":"ContainerStarted","Data":"0953ee3930b7da1187ed32b45a8e3a8a7c45af969af4545f8327d668d2852f6a"} Mar 18 00:33:41 crc kubenswrapper[4983]: I0318 00:33:41.194983 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" event={"ID":"3753b92e-109f-4875-a1fe-70ea96a52a29","Type":"ContainerStarted","Data":"90ee01951ce8cc5ce982c7ad64f45ea2f988586f53eb3568b66d5a576a10982d"} Mar 18 00:33:41 crc kubenswrapper[4983]: I0318 00:33:41.201401 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"6c1ecce7-261c-4f82-863c-641ef76200ae","Type":"ContainerStarted","Data":"d34e480cf85e8ec9bf588aec252d7e03242aa5054520d3675d7dc5657de7430d"} Mar 18 00:33:41 crc kubenswrapper[4983]: I0318 00:33:41.201441 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"6c1ecce7-261c-4f82-863c-641ef76200ae","Type":"ContainerStarted","Data":"3733349a29810d16fb4e4ea45cda6a0d47521cb107033512cb3f2a4d6d7dc9c5"} Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.601385 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/alertmanager-default-0" podStartSLOduration=16.62826946 podStartE2EDuration="27.601362964s" podCreationTimestamp="2026-03-18 00:33:15 +0000 UTC" firstStartedPulling="2026-03-18 00:33:30.08650845 +0000 UTC m=+1527.983883060" lastFinishedPulling="2026-03-18 00:33:41.059601954 +0000 UTC m=+1538.956976564" observedRunningTime="2026-03-18 00:33:41.225667275 +0000 UTC m=+1539.123041905" watchObservedRunningTime="2026-03-18 00:33:42.601362964 +0000 UTC m=+1540.498737574" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.608040 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2"] Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.609090 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.610865 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-event-sg-core-configmap" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.611084 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-cert" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.622835 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2"] Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.711737 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg8d2\" (UniqueName: \"kubernetes.io/projected/2ba80112-3269-42af-a92b-86876fe20c01-kube-api-access-qg8d2\") pod \"default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2\" (UID: \"2ba80112-3269-42af-a92b-86876fe20c01\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.711848 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/2ba80112-3269-42af-a92b-86876fe20c01-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2\" (UID: \"2ba80112-3269-42af-a92b-86876fe20c01\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.711886 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/2ba80112-3269-42af-a92b-86876fe20c01-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2\" (UID: \"2ba80112-3269-42af-a92b-86876fe20c01\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.711953 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/2ba80112-3269-42af-a92b-86876fe20c01-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2\" (UID: \"2ba80112-3269-42af-a92b-86876fe20c01\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.813432 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/2ba80112-3269-42af-a92b-86876fe20c01-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2\" (UID: \"2ba80112-3269-42af-a92b-86876fe20c01\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.813517 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg8d2\" (UniqueName: \"kubernetes.io/projected/2ba80112-3269-42af-a92b-86876fe20c01-kube-api-access-qg8d2\") pod \"default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2\" (UID: \"2ba80112-3269-42af-a92b-86876fe20c01\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.813598 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/2ba80112-3269-42af-a92b-86876fe20c01-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2\" (UID: \"2ba80112-3269-42af-a92b-86876fe20c01\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.813633 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/2ba80112-3269-42af-a92b-86876fe20c01-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2\" (UID: \"2ba80112-3269-42af-a92b-86876fe20c01\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.814499 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/2ba80112-3269-42af-a92b-86876fe20c01-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2\" (UID: \"2ba80112-3269-42af-a92b-86876fe20c01\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.815054 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/2ba80112-3269-42af-a92b-86876fe20c01-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2\" (UID: \"2ba80112-3269-42af-a92b-86876fe20c01\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.819678 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/2ba80112-3269-42af-a92b-86876fe20c01-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2\" (UID: \"2ba80112-3269-42af-a92b-86876fe20c01\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.834251 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg8d2\" (UniqueName: \"kubernetes.io/projected/2ba80112-3269-42af-a92b-86876fe20c01-kube-api-access-qg8d2\") pod \"default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2\" (UID: \"2ba80112-3269-42af-a92b-86876fe20c01\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:42 crc kubenswrapper[4983]: I0318 00:33:42.949661 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" Mar 18 00:33:43 crc kubenswrapper[4983]: I0318 00:33:43.259595 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:33:43 crc kubenswrapper[4983]: I0318 00:33:43.259647 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:33:43 crc kubenswrapper[4983]: I0318 00:33:43.699122 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2"] Mar 18 00:33:43 crc kubenswrapper[4983]: W0318 00:33:43.716506 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ba80112_3269_42af_a92b_86876fe20c01.slice/crio-5aa521f9fe320e6e6de65f18ff6b9b413b5f48b99d467c6eebebf4c0df490739 WatchSource:0}: Error finding container 5aa521f9fe320e6e6de65f18ff6b9b413b5f48b99d467c6eebebf4c0df490739: Status 404 returned error can't find the container with id 5aa521f9fe320e6e6de65f18ff6b9b413b5f48b99d467c6eebebf4c0df490739 Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.145482 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g"] Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.148724 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.153367 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-event-sg-core-configmap" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.156142 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g"] Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.228851 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" event={"ID":"937b4985-05d5-48c0-b408-3164619e74ca","Type":"ContainerStarted","Data":"f449ada84607f67ed1fbedb2c9f67426e9f022f1cf26756b6ff3d8d7b0c93ac7"} Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.231279 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" event={"ID":"3753b92e-109f-4875-a1fe-70ea96a52a29","Type":"ContainerStarted","Data":"5a957d60d5de59a77e306194d841252fee69843816bbeb5009aa3fa84b7dc6c2"} Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.234910 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" event={"ID":"2ba80112-3269-42af-a92b-86876fe20c01","Type":"ContainerStarted","Data":"446b7a9de5f323f22af99567d18bcbdb310c9463fe26ede8ef649605a83856b0"} Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.234967 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" event={"ID":"2ba80112-3269-42af-a92b-86876fe20c01","Type":"ContainerStarted","Data":"1430b2b293251fa7db0e99754dc43ec861a2f6a0f4f21bcce7c4e7711bbd3e1c"} Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.234982 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" event={"ID":"2ba80112-3269-42af-a92b-86876fe20c01","Type":"ContainerStarted","Data":"5aa521f9fe320e6e6de65f18ff6b9b413b5f48b99d467c6eebebf4c0df490739"} Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.237063 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" event={"ID":"65a13175-2c14-45b7-92f5-4677c93be37f","Type":"ContainerStarted","Data":"ddff2e5be9085e317c98fe2984d3e9cd46920a5c490e4711961c2570f6ddca33"} Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.244151 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/fcaaaeb3-eae9-44ee-8fc9-348df5c997a9-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g\" (UID: \"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.244247 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/fcaaaeb3-eae9-44ee-8fc9-348df5c997a9-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g\" (UID: \"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.244276 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/fcaaaeb3-eae9-44ee-8fc9-348df5c997a9-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g\" (UID: \"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.244344 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf7xx\" (UniqueName: \"kubernetes.io/projected/fcaaaeb3-eae9-44ee-8fc9-348df5c997a9-kube-api-access-bf7xx\") pod \"default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g\" (UID: \"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.249889 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" podStartSLOduration=6.809577395 podStartE2EDuration="12.249876212s" podCreationTimestamp="2026-03-18 00:33:32 +0000 UTC" firstStartedPulling="2026-03-18 00:33:38.116778956 +0000 UTC m=+1536.014153566" lastFinishedPulling="2026-03-18 00:33:43.557077773 +0000 UTC m=+1541.454452383" observedRunningTime="2026-03-18 00:33:44.245210132 +0000 UTC m=+1542.142584752" watchObservedRunningTime="2026-03-18 00:33:44.249876212 +0000 UTC m=+1542.147250822" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.286339 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" podStartSLOduration=4.249328527 podStartE2EDuration="9.286309808s" podCreationTimestamp="2026-03-18 00:33:35 +0000 UTC" firstStartedPulling="2026-03-18 00:33:38.520875334 +0000 UTC m=+1536.418249944" lastFinishedPulling="2026-03-18 00:33:43.557856615 +0000 UTC m=+1541.455231225" observedRunningTime="2026-03-18 00:33:44.266291159 +0000 UTC m=+1542.163665779" watchObservedRunningTime="2026-03-18 00:33:44.286309808 +0000 UTC m=+1542.183684428" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.306682 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" podStartSLOduration=8.528922204 podStartE2EDuration="15.306655255s" podCreationTimestamp="2026-03-18 00:33:29 +0000 UTC" firstStartedPulling="2026-03-18 00:33:36.817122246 +0000 UTC m=+1534.714496866" lastFinishedPulling="2026-03-18 00:33:43.594855307 +0000 UTC m=+1541.492229917" observedRunningTime="2026-03-18 00:33:44.29753068 +0000 UTC m=+1542.194905300" watchObservedRunningTime="2026-03-18 00:33:44.306655255 +0000 UTC m=+1542.204029865" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.330954 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" podStartSLOduration=2.035197225 podStartE2EDuration="2.330912041s" podCreationTimestamp="2026-03-18 00:33:42 +0000 UTC" firstStartedPulling="2026-03-18 00:33:43.720369667 +0000 UTC m=+1541.617744277" lastFinishedPulling="2026-03-18 00:33:44.016084483 +0000 UTC m=+1541.913459093" observedRunningTime="2026-03-18 00:33:44.329358318 +0000 UTC m=+1542.226732928" watchObservedRunningTime="2026-03-18 00:33:44.330912041 +0000 UTC m=+1542.228286651" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.345320 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf7xx\" (UniqueName: \"kubernetes.io/projected/fcaaaeb3-eae9-44ee-8fc9-348df5c997a9-kube-api-access-bf7xx\") pod \"default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g\" (UID: \"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.345442 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/fcaaaeb3-eae9-44ee-8fc9-348df5c997a9-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g\" (UID: \"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.345538 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/fcaaaeb3-eae9-44ee-8fc9-348df5c997a9-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g\" (UID: \"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.345582 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/fcaaaeb3-eae9-44ee-8fc9-348df5c997a9-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g\" (UID: \"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.346784 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/fcaaaeb3-eae9-44ee-8fc9-348df5c997a9-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g\" (UID: \"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.350098 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/fcaaaeb3-eae9-44ee-8fc9-348df5c997a9-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g\" (UID: \"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.368873 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf7xx\" (UniqueName: \"kubernetes.io/projected/fcaaaeb3-eae9-44ee-8fc9-348df5c997a9-kube-api-access-bf7xx\") pod \"default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g\" (UID: \"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.368916 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/fcaaaeb3-eae9-44ee-8fc9-348df5c997a9-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g\" (UID: \"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.471564 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" Mar 18 00:33:44 crc kubenswrapper[4983]: I0318 00:33:44.884242 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g"] Mar 18 00:33:45 crc kubenswrapper[4983]: I0318 00:33:45.258556 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" event={"ID":"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9","Type":"ContainerStarted","Data":"88591442c0921553721e02040908e840c7a8fb3c2935bddcb301b007fa9495b8"} Mar 18 00:33:45 crc kubenswrapper[4983]: I0318 00:33:45.258840 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" event={"ID":"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9","Type":"ContainerStarted","Data":"00405444c4b7d06f79acb34c3f34e045997d973c0578e25802d7f4b7c3a244de"} Mar 18 00:33:46 crc kubenswrapper[4983]: I0318 00:33:46.268059 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" event={"ID":"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9","Type":"ContainerStarted","Data":"4c8b39fd6f84c4018f7960a98d3b9026a66b6ae09e445d3aaaaf93bb94e71d6c"} Mar 18 00:33:46 crc kubenswrapper[4983]: I0318 00:33:46.287968 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" podStartSLOduration=1.9616514619999998 podStartE2EDuration="2.28792458s" podCreationTimestamp="2026-03-18 00:33:44 +0000 UTC" firstStartedPulling="2026-03-18 00:33:44.895410512 +0000 UTC m=+1542.792785132" lastFinishedPulling="2026-03-18 00:33:45.22168364 +0000 UTC m=+1543.119058250" observedRunningTime="2026-03-18 00:33:46.28719051 +0000 UTC m=+1544.184565120" watchObservedRunningTime="2026-03-18 00:33:46.28792458 +0000 UTC m=+1544.185299200" Mar 18 00:33:49 crc kubenswrapper[4983]: I0318 00:33:49.313781 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/prometheus-default-0" Mar 18 00:33:49 crc kubenswrapper[4983]: I0318 00:33:49.367134 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/prometheus-default-0" Mar 18 00:33:50 crc kubenswrapper[4983]: I0318 00:33:50.338039 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/prometheus-default-0" Mar 18 00:33:54 crc kubenswrapper[4983]: I0318 00:33:54.878457 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-2rwvf"] Mar 18 00:33:54 crc kubenswrapper[4983]: I0318 00:33:54.879246 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" podUID="b26927e3-2404-4f34-9fd5-afb37a1f4387" containerName="default-interconnect" containerID="cri-o://47a42d8b569bee2ab9848da5a9e7e5a310cd2380de8e205a4eeccb5220929297" gracePeriod=30 Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.323542 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.333892 4983 generic.go:334] "Generic (PLEG): container finished" podID="2ba80112-3269-42af-a92b-86876fe20c01" containerID="1430b2b293251fa7db0e99754dc43ec861a2f6a0f4f21bcce7c4e7711bbd3e1c" exitCode=0 Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.333965 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" event={"ID":"2ba80112-3269-42af-a92b-86876fe20c01","Type":"ContainerDied","Data":"1430b2b293251fa7db0e99754dc43ec861a2f6a0f4f21bcce7c4e7711bbd3e1c"} Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.334882 4983 scope.go:117] "RemoveContainer" containerID="1430b2b293251fa7db0e99754dc43ec861a2f6a0f4f21bcce7c4e7711bbd3e1c" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.335587 4983 generic.go:334] "Generic (PLEG): container finished" podID="b26927e3-2404-4f34-9fd5-afb37a1f4387" containerID="47a42d8b569bee2ab9848da5a9e7e5a310cd2380de8e205a4eeccb5220929297" exitCode=0 Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.335646 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.335678 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" event={"ID":"b26927e3-2404-4f34-9fd5-afb37a1f4387","Type":"ContainerDied","Data":"47a42d8b569bee2ab9848da5a9e7e5a310cd2380de8e205a4eeccb5220929297"} Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.335748 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-2rwvf" event={"ID":"b26927e3-2404-4f34-9fd5-afb37a1f4387","Type":"ContainerDied","Data":"fd8d6f4b385a812b39f14b9a35923d5aebd703130084db58a0918a3f957c54ee"} Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.335774 4983 scope.go:117] "RemoveContainer" containerID="47a42d8b569bee2ab9848da5a9e7e5a310cd2380de8e205a4eeccb5220929297" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.338287 4983 generic.go:334] "Generic (PLEG): container finished" podID="fcaaaeb3-eae9-44ee-8fc9-348df5c997a9" containerID="88591442c0921553721e02040908e840c7a8fb3c2935bddcb301b007fa9495b8" exitCode=0 Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.338423 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" event={"ID":"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9","Type":"ContainerDied","Data":"88591442c0921553721e02040908e840c7a8fb3c2935bddcb301b007fa9495b8"} Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.338900 4983 scope.go:117] "RemoveContainer" containerID="88591442c0921553721e02040908e840c7a8fb3c2935bddcb301b007fa9495b8" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.351092 4983 generic.go:334] "Generic (PLEG): container finished" podID="65a13175-2c14-45b7-92f5-4677c93be37f" containerID="0953ee3930b7da1187ed32b45a8e3a8a7c45af969af4545f8327d668d2852f6a" exitCode=0 Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.351148 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" event={"ID":"65a13175-2c14-45b7-92f5-4677c93be37f","Type":"ContainerDied","Data":"0953ee3930b7da1187ed32b45a8e3a8a7c45af969af4545f8327d668d2852f6a"} Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.352012 4983 scope.go:117] "RemoveContainer" containerID="0953ee3930b7da1187ed32b45a8e3a8a7c45af969af4545f8327d668d2852f6a" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.364945 4983 scope.go:117] "RemoveContainer" containerID="47a42d8b569bee2ab9848da5a9e7e5a310cd2380de8e205a4eeccb5220929297" Mar 18 00:33:55 crc kubenswrapper[4983]: E0318 00:33:55.366578 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47a42d8b569bee2ab9848da5a9e7e5a310cd2380de8e205a4eeccb5220929297\": container with ID starting with 47a42d8b569bee2ab9848da5a9e7e5a310cd2380de8e205a4eeccb5220929297 not found: ID does not exist" containerID="47a42d8b569bee2ab9848da5a9e7e5a310cd2380de8e205a4eeccb5220929297" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.366619 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47a42d8b569bee2ab9848da5a9e7e5a310cd2380de8e205a4eeccb5220929297"} err="failed to get container status \"47a42d8b569bee2ab9848da5a9e7e5a310cd2380de8e205a4eeccb5220929297\": rpc error: code = NotFound desc = could not find container \"47a42d8b569bee2ab9848da5a9e7e5a310cd2380de8e205a4eeccb5220929297\": container with ID starting with 47a42d8b569bee2ab9848da5a9e7e5a310cd2380de8e205a4eeccb5220929297 not found: ID does not exist" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.507715 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/b26927e3-2404-4f34-9fd5-afb37a1f4387-sasl-config\") pod \"b26927e3-2404-4f34-9fd5-afb37a1f4387\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.508191 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-openstack-ca\") pod \"b26927e3-2404-4f34-9fd5-afb37a1f4387\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.508233 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-inter-router-ca\") pod \"b26927e3-2404-4f34-9fd5-afb37a1f4387\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.508267 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-openstack-credentials\") pod \"b26927e3-2404-4f34-9fd5-afb37a1f4387\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.508363 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-inter-router-credentials\") pod \"b26927e3-2404-4f34-9fd5-afb37a1f4387\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.508403 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-sasl-users\") pod \"b26927e3-2404-4f34-9fd5-afb37a1f4387\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.508484 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwr7k\" (UniqueName: \"kubernetes.io/projected/b26927e3-2404-4f34-9fd5-afb37a1f4387-kube-api-access-cwr7k\") pod \"b26927e3-2404-4f34-9fd5-afb37a1f4387\" (UID: \"b26927e3-2404-4f34-9fd5-afb37a1f4387\") " Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.509689 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b26927e3-2404-4f34-9fd5-afb37a1f4387-sasl-config" (OuterVolumeSpecName: "sasl-config") pod "b26927e3-2404-4f34-9fd5-afb37a1f4387" (UID: "b26927e3-2404-4f34-9fd5-afb37a1f4387"). InnerVolumeSpecName "sasl-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.514820 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-openstack-credentials" (OuterVolumeSpecName: "default-interconnect-openstack-credentials") pod "b26927e3-2404-4f34-9fd5-afb37a1f4387" (UID: "b26927e3-2404-4f34-9fd5-afb37a1f4387"). InnerVolumeSpecName "default-interconnect-openstack-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.515421 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-sasl-users" (OuterVolumeSpecName: "sasl-users") pod "b26927e3-2404-4f34-9fd5-afb37a1f4387" (UID: "b26927e3-2404-4f34-9fd5-afb37a1f4387"). InnerVolumeSpecName "sasl-users". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.516810 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-inter-router-ca" (OuterVolumeSpecName: "default-interconnect-inter-router-ca") pod "b26927e3-2404-4f34-9fd5-afb37a1f4387" (UID: "b26927e3-2404-4f34-9fd5-afb37a1f4387"). InnerVolumeSpecName "default-interconnect-inter-router-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.518232 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-openstack-ca" (OuterVolumeSpecName: "default-interconnect-openstack-ca") pod "b26927e3-2404-4f34-9fd5-afb37a1f4387" (UID: "b26927e3-2404-4f34-9fd5-afb37a1f4387"). InnerVolumeSpecName "default-interconnect-openstack-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.521219 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b26927e3-2404-4f34-9fd5-afb37a1f4387-kube-api-access-cwr7k" (OuterVolumeSpecName: "kube-api-access-cwr7k") pod "b26927e3-2404-4f34-9fd5-afb37a1f4387" (UID: "b26927e3-2404-4f34-9fd5-afb37a1f4387"). InnerVolumeSpecName "kube-api-access-cwr7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.525291 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-inter-router-credentials" (OuterVolumeSpecName: "default-interconnect-inter-router-credentials") pod "b26927e3-2404-4f34-9fd5-afb37a1f4387" (UID: "b26927e3-2404-4f34-9fd5-afb37a1f4387"). InnerVolumeSpecName "default-interconnect-inter-router-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.610870 4983 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-inter-router-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.610999 4983 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-openstack-credentials\") on node \"crc\" DevicePath \"\"" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.611061 4983 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-inter-router-credentials\") on node \"crc\" DevicePath \"\"" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.611133 4983 reconciler_common.go:293] "Volume detached for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-sasl-users\") on node \"crc\" DevicePath \"\"" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.611191 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwr7k\" (UniqueName: \"kubernetes.io/projected/b26927e3-2404-4f34-9fd5-afb37a1f4387-kube-api-access-cwr7k\") on node \"crc\" DevicePath \"\"" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.611259 4983 reconciler_common.go:293] "Volume detached for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/b26927e3-2404-4f34-9fd5-afb37a1f4387-sasl-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.611321 4983 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/b26927e3-2404-4f34-9fd5-afb37a1f4387-default-interconnect-openstack-ca\") on node \"crc\" DevicePath \"\"" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.690195 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-2rwvf"] Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.713414 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-2rwvf"] Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.919862 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-jn22d"] Mar 18 00:33:55 crc kubenswrapper[4983]: E0318 00:33:55.920212 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b26927e3-2404-4f34-9fd5-afb37a1f4387" containerName="default-interconnect" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.920225 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="b26927e3-2404-4f34-9fd5-afb37a1f4387" containerName="default-interconnect" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.920351 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="b26927e3-2404-4f34-9fd5-afb37a1f4387" containerName="default-interconnect" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.920817 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.928398 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.929834 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.930393 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-hgsq6" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.930516 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.930529 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.931940 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.932906 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Mar 18 00:33:55 crc kubenswrapper[4983]: I0318 00:33:55.937166 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-jn22d"] Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.020446 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.020572 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.020628 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r79sx\" (UniqueName: \"kubernetes.io/projected/194c9fd4-aedf-4d03-a82d-077903e33e44-kube-api-access-r79sx\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.020662 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.020684 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/194c9fd4-aedf-4d03-a82d-077903e33e44-sasl-config\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.020787 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-sasl-users\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.020887 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.121513 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.121625 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.121659 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r79sx\" (UniqueName: \"kubernetes.io/projected/194c9fd4-aedf-4d03-a82d-077903e33e44-kube-api-access-r79sx\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.121692 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.121716 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/194c9fd4-aedf-4d03-a82d-077903e33e44-sasl-config\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.121759 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-sasl-users\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.121788 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.123014 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/194c9fd4-aedf-4d03-a82d-077903e33e44-sasl-config\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.128102 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.129760 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-sasl-users\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.130169 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.136028 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.138859 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/194c9fd4-aedf-4d03-a82d-077903e33e44-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.145818 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r79sx\" (UniqueName: \"kubernetes.io/projected/194c9fd4-aedf-4d03-a82d-077903e33e44-kube-api-access-r79sx\") pod \"default-interconnect-68864d46cb-jn22d\" (UID: \"194c9fd4-aedf-4d03-a82d-077903e33e44\") " pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.242384 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-jn22d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.367998 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" event={"ID":"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9","Type":"ContainerStarted","Data":"d07e637f0508f4034d3a16a86642f2c142b78eed7963870c276506e93bc14bd5"} Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.372546 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" event={"ID":"65a13175-2c14-45b7-92f5-4677c93be37f","Type":"ContainerStarted","Data":"bd0392b28bda01f0b79a3679f23a700610d020a81d8cd8d02978d3716ea179b7"} Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.378152 4983 generic.go:334] "Generic (PLEG): container finished" podID="937b4985-05d5-48c0-b408-3164619e74ca" containerID="ab45076cc85114c7c2402083464275316cc1881b0644bf60a56a6132ec1a1ad0" exitCode=0 Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.378211 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" event={"ID":"937b4985-05d5-48c0-b408-3164619e74ca","Type":"ContainerDied","Data":"ab45076cc85114c7c2402083464275316cc1881b0644bf60a56a6132ec1a1ad0"} Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.378676 4983 scope.go:117] "RemoveContainer" containerID="ab45076cc85114c7c2402083464275316cc1881b0644bf60a56a6132ec1a1ad0" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.381191 4983 generic.go:334] "Generic (PLEG): container finished" podID="3753b92e-109f-4875-a1fe-70ea96a52a29" containerID="90ee01951ce8cc5ce982c7ad64f45ea2f988586f53eb3568b66d5a576a10982d" exitCode=0 Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.381255 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" event={"ID":"3753b92e-109f-4875-a1fe-70ea96a52a29","Type":"ContainerDied","Data":"90ee01951ce8cc5ce982c7ad64f45ea2f988586f53eb3568b66d5a576a10982d"} Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.381590 4983 scope.go:117] "RemoveContainer" containerID="90ee01951ce8cc5ce982c7ad64f45ea2f988586f53eb3568b66d5a576a10982d" Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.384641 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" event={"ID":"2ba80112-3269-42af-a92b-86876fe20c01","Type":"ContainerStarted","Data":"64f6fb2c5a225889892516fb2259bd64047b795f6a443c063da26ff7cc7e305e"} Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.657759 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-jn22d"] Mar 18 00:33:56 crc kubenswrapper[4983]: I0318 00:33:56.852523 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b26927e3-2404-4f34-9fd5-afb37a1f4387" path="/var/lib/kubelet/pods/b26927e3-2404-4f34-9fd5-afb37a1f4387/volumes" Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.396319 4983 generic.go:334] "Generic (PLEG): container finished" podID="fcaaaeb3-eae9-44ee-8fc9-348df5c997a9" containerID="d07e637f0508f4034d3a16a86642f2c142b78eed7963870c276506e93bc14bd5" exitCode=0 Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.396406 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" event={"ID":"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9","Type":"ContainerDied","Data":"d07e637f0508f4034d3a16a86642f2c142b78eed7963870c276506e93bc14bd5"} Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.396561 4983 scope.go:117] "RemoveContainer" containerID="88591442c0921553721e02040908e840c7a8fb3c2935bddcb301b007fa9495b8" Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.397316 4983 scope.go:117] "RemoveContainer" containerID="d07e637f0508f4034d3a16a86642f2c142b78eed7963870c276506e93bc14bd5" Mar 18 00:33:57 crc kubenswrapper[4983]: E0318 00:33:57.397693 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g_service-telemetry(fcaaaeb3-eae9-44ee-8fc9-348df5c997a9)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" podUID="fcaaaeb3-eae9-44ee-8fc9-348df5c997a9" Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.401194 4983 generic.go:334] "Generic (PLEG): container finished" podID="65a13175-2c14-45b7-92f5-4677c93be37f" containerID="bd0392b28bda01f0b79a3679f23a700610d020a81d8cd8d02978d3716ea179b7" exitCode=0 Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.401276 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" event={"ID":"65a13175-2c14-45b7-92f5-4677c93be37f","Type":"ContainerDied","Data":"bd0392b28bda01f0b79a3679f23a700610d020a81d8cd8d02978d3716ea179b7"} Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.401892 4983 scope.go:117] "RemoveContainer" containerID="bd0392b28bda01f0b79a3679f23a700610d020a81d8cd8d02978d3716ea179b7" Mar 18 00:33:57 crc kubenswrapper[4983]: E0318 00:33:57.402122 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq_service-telemetry(65a13175-2c14-45b7-92f5-4677c93be37f)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" podUID="65a13175-2c14-45b7-92f5-4677c93be37f" Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.404916 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" event={"ID":"937b4985-05d5-48c0-b408-3164619e74ca","Type":"ContainerStarted","Data":"ee6750b5c3ce747e6604f74fb7b704f65681d582eab3c6429e39e89c6be1b546"} Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.407730 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" event={"ID":"3753b92e-109f-4875-a1fe-70ea96a52a29","Type":"ContainerStarted","Data":"7be6a908d8aea62e2c9ee64bf8187c8c256f998e60416d9e03b4b2b349ca9e6b"} Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.411047 4983 generic.go:334] "Generic (PLEG): container finished" podID="2ba80112-3269-42af-a92b-86876fe20c01" containerID="64f6fb2c5a225889892516fb2259bd64047b795f6a443c063da26ff7cc7e305e" exitCode=0 Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.411158 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" event={"ID":"2ba80112-3269-42af-a92b-86876fe20c01","Type":"ContainerDied","Data":"64f6fb2c5a225889892516fb2259bd64047b795f6a443c063da26ff7cc7e305e"} Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.411736 4983 scope.go:117] "RemoveContainer" containerID="64f6fb2c5a225889892516fb2259bd64047b795f6a443c063da26ff7cc7e305e" Mar 18 00:33:57 crc kubenswrapper[4983]: E0318 00:33:57.412031 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2_service-telemetry(2ba80112-3269-42af-a92b-86876fe20c01)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" podUID="2ba80112-3269-42af-a92b-86876fe20c01" Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.412744 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-jn22d" event={"ID":"194c9fd4-aedf-4d03-a82d-077903e33e44","Type":"ContainerStarted","Data":"4d5dd24a5c1f8e5f563680bcfd3f3e32c2ecaa9ee5d4f6da17c2585c75f2f5ae"} Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.412782 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-jn22d" event={"ID":"194c9fd4-aedf-4d03-a82d-077903e33e44","Type":"ContainerStarted","Data":"593bf91b743cc99f8e3abbcd4bc515ee6c3dfcbd9a022be7ac8bf8860063c947"} Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.467181 4983 scope.go:117] "RemoveContainer" containerID="0953ee3930b7da1187ed32b45a8e3a8a7c45af969af4545f8327d668d2852f6a" Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.537438 4983 scope.go:117] "RemoveContainer" containerID="1430b2b293251fa7db0e99754dc43ec861a2f6a0f4f21bcce7c4e7711bbd3e1c" Mar 18 00:33:57 crc kubenswrapper[4983]: I0318 00:33:57.538149 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-jn22d" podStartSLOduration=3.538131603 podStartE2EDuration="3.538131603s" podCreationTimestamp="2026-03-18 00:33:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 00:33:57.537771283 +0000 UTC m=+1555.435145903" watchObservedRunningTime="2026-03-18 00:33:57.538131603 +0000 UTC m=+1555.435506203" Mar 18 00:33:58 crc kubenswrapper[4983]: I0318 00:33:58.428490 4983 generic.go:334] "Generic (PLEG): container finished" podID="937b4985-05d5-48c0-b408-3164619e74ca" containerID="ee6750b5c3ce747e6604f74fb7b704f65681d582eab3c6429e39e89c6be1b546" exitCode=0 Mar 18 00:33:58 crc kubenswrapper[4983]: I0318 00:33:58.428570 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" event={"ID":"937b4985-05d5-48c0-b408-3164619e74ca","Type":"ContainerDied","Data":"ee6750b5c3ce747e6604f74fb7b704f65681d582eab3c6429e39e89c6be1b546"} Mar 18 00:33:58 crc kubenswrapper[4983]: I0318 00:33:58.428954 4983 scope.go:117] "RemoveContainer" containerID="ab45076cc85114c7c2402083464275316cc1881b0644bf60a56a6132ec1a1ad0" Mar 18 00:33:58 crc kubenswrapper[4983]: I0318 00:33:58.429428 4983 scope.go:117] "RemoveContainer" containerID="ee6750b5c3ce747e6604f74fb7b704f65681d582eab3c6429e39e89c6be1b546" Mar 18 00:33:58 crc kubenswrapper[4983]: E0318 00:33:58.429622 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7_service-telemetry(937b4985-05d5-48c0-b408-3164619e74ca)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" podUID="937b4985-05d5-48c0-b408-3164619e74ca" Mar 18 00:33:58 crc kubenswrapper[4983]: I0318 00:33:58.434181 4983 generic.go:334] "Generic (PLEG): container finished" podID="3753b92e-109f-4875-a1fe-70ea96a52a29" containerID="7be6a908d8aea62e2c9ee64bf8187c8c256f998e60416d9e03b4b2b349ca9e6b" exitCode=0 Mar 18 00:33:58 crc kubenswrapper[4983]: I0318 00:33:58.434396 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" event={"ID":"3753b92e-109f-4875-a1fe-70ea96a52a29","Type":"ContainerDied","Data":"7be6a908d8aea62e2c9ee64bf8187c8c256f998e60416d9e03b4b2b349ca9e6b"} Mar 18 00:33:58 crc kubenswrapper[4983]: I0318 00:33:58.435394 4983 scope.go:117] "RemoveContainer" containerID="7be6a908d8aea62e2c9ee64bf8187c8c256f998e60416d9e03b4b2b349ca9e6b" Mar 18 00:33:58 crc kubenswrapper[4983]: E0318 00:33:58.436072 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp_service-telemetry(3753b92e-109f-4875-a1fe-70ea96a52a29)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" podUID="3753b92e-109f-4875-a1fe-70ea96a52a29" Mar 18 00:33:58 crc kubenswrapper[4983]: I0318 00:33:58.470120 4983 scope.go:117] "RemoveContainer" containerID="90ee01951ce8cc5ce982c7ad64f45ea2f988586f53eb3568b66d5a576a10982d" Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.271181 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/qdr-test"] Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.271969 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.274021 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"qdr-test-config" Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.274210 4983 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-selfsigned" Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.290855 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.381203 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/c0983273-84f3-4f26-bef5-e738d038c3d6-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"c0983273-84f3-4f26-bef5-e738d038c3d6\") " pod="service-telemetry/qdr-test" Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.381265 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfktj\" (UniqueName: \"kubernetes.io/projected/c0983273-84f3-4f26-bef5-e738d038c3d6-kube-api-access-vfktj\") pod \"qdr-test\" (UID: \"c0983273-84f3-4f26-bef5-e738d038c3d6\") " pod="service-telemetry/qdr-test" Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.381330 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/c0983273-84f3-4f26-bef5-e738d038c3d6-qdr-test-config\") pod \"qdr-test\" (UID: \"c0983273-84f3-4f26-bef5-e738d038c3d6\") " pod="service-telemetry/qdr-test" Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.483108 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/c0983273-84f3-4f26-bef5-e738d038c3d6-qdr-test-config\") pod \"qdr-test\" (UID: \"c0983273-84f3-4f26-bef5-e738d038c3d6\") " pod="service-telemetry/qdr-test" Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.483208 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/c0983273-84f3-4f26-bef5-e738d038c3d6-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"c0983273-84f3-4f26-bef5-e738d038c3d6\") " pod="service-telemetry/qdr-test" Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.483253 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfktj\" (UniqueName: \"kubernetes.io/projected/c0983273-84f3-4f26-bef5-e738d038c3d6-kube-api-access-vfktj\") pod \"qdr-test\" (UID: \"c0983273-84f3-4f26-bef5-e738d038c3d6\") " pod="service-telemetry/qdr-test" Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.484303 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/c0983273-84f3-4f26-bef5-e738d038c3d6-qdr-test-config\") pod \"qdr-test\" (UID: \"c0983273-84f3-4f26-bef5-e738d038c3d6\") " pod="service-telemetry/qdr-test" Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.496788 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/c0983273-84f3-4f26-bef5-e738d038c3d6-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"c0983273-84f3-4f26-bef5-e738d038c3d6\") " pod="service-telemetry/qdr-test" Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.501510 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfktj\" (UniqueName: \"kubernetes.io/projected/c0983273-84f3-4f26-bef5-e738d038c3d6-kube-api-access-vfktj\") pod \"qdr-test\" (UID: \"c0983273-84f3-4f26-bef5-e738d038c3d6\") " pod="service-telemetry/qdr-test" Mar 18 00:33:59 crc kubenswrapper[4983]: I0318 00:33:59.587845 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Mar 18 00:34:00 crc kubenswrapper[4983]: I0318 00:34:00.018745 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Mar 18 00:34:00 crc kubenswrapper[4983]: W0318 00:34:00.034787 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0983273_84f3_4f26_bef5_e738d038c3d6.slice/crio-06b44fce90372fa0b18e4d209aa18c77103726241e432179421e71dead3e4ef6 WatchSource:0}: Error finding container 06b44fce90372fa0b18e4d209aa18c77103726241e432179421e71dead3e4ef6: Status 404 returned error can't find the container with id 06b44fce90372fa0b18e4d209aa18c77103726241e432179421e71dead3e4ef6 Mar 18 00:34:00 crc kubenswrapper[4983]: I0318 00:34:00.137495 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563234-pzlvr"] Mar 18 00:34:00 crc kubenswrapper[4983]: I0318 00:34:00.138465 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563234-pzlvr" Mar 18 00:34:00 crc kubenswrapper[4983]: I0318 00:34:00.141808 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563234-pzlvr"] Mar 18 00:34:00 crc kubenswrapper[4983]: I0318 00:34:00.144180 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:34:00 crc kubenswrapper[4983]: I0318 00:34:00.144523 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:34:00 crc kubenswrapper[4983]: I0318 00:34:00.144709 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:34:00 crc kubenswrapper[4983]: I0318 00:34:00.294303 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d7zf\" (UniqueName: \"kubernetes.io/projected/5f02470c-1df4-4c79-b670-460dcecb6c68-kube-api-access-8d7zf\") pod \"auto-csr-approver-29563234-pzlvr\" (UID: \"5f02470c-1df4-4c79-b670-460dcecb6c68\") " pod="openshift-infra/auto-csr-approver-29563234-pzlvr" Mar 18 00:34:00 crc kubenswrapper[4983]: I0318 00:34:00.395771 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d7zf\" (UniqueName: \"kubernetes.io/projected/5f02470c-1df4-4c79-b670-460dcecb6c68-kube-api-access-8d7zf\") pod \"auto-csr-approver-29563234-pzlvr\" (UID: \"5f02470c-1df4-4c79-b670-460dcecb6c68\") " pod="openshift-infra/auto-csr-approver-29563234-pzlvr" Mar 18 00:34:00 crc kubenswrapper[4983]: I0318 00:34:00.417596 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d7zf\" (UniqueName: \"kubernetes.io/projected/5f02470c-1df4-4c79-b670-460dcecb6c68-kube-api-access-8d7zf\") pod \"auto-csr-approver-29563234-pzlvr\" (UID: \"5f02470c-1df4-4c79-b670-460dcecb6c68\") " pod="openshift-infra/auto-csr-approver-29563234-pzlvr" Mar 18 00:34:00 crc kubenswrapper[4983]: I0318 00:34:00.455000 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"c0983273-84f3-4f26-bef5-e738d038c3d6","Type":"ContainerStarted","Data":"06b44fce90372fa0b18e4d209aa18c77103726241e432179421e71dead3e4ef6"} Mar 18 00:34:00 crc kubenswrapper[4983]: I0318 00:34:00.481009 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563234-pzlvr" Mar 18 00:34:00 crc kubenswrapper[4983]: I0318 00:34:00.708690 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563234-pzlvr"] Mar 18 00:34:00 crc kubenswrapper[4983]: W0318 00:34:00.716480 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f02470c_1df4_4c79_b670_460dcecb6c68.slice/crio-09c6d163e139400f12ee181ab5c88d790fc8718b23099b41b6fcff517e80629f WatchSource:0}: Error finding container 09c6d163e139400f12ee181ab5c88d790fc8718b23099b41b6fcff517e80629f: Status 404 returned error can't find the container with id 09c6d163e139400f12ee181ab5c88d790fc8718b23099b41b6fcff517e80629f Mar 18 00:34:01 crc kubenswrapper[4983]: I0318 00:34:01.467696 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563234-pzlvr" event={"ID":"5f02470c-1df4-4c79-b670-460dcecb6c68","Type":"ContainerStarted","Data":"09c6d163e139400f12ee181ab5c88d790fc8718b23099b41b6fcff517e80629f"} Mar 18 00:34:02 crc kubenswrapper[4983]: I0318 00:34:02.476696 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563234-pzlvr" event={"ID":"5f02470c-1df4-4c79-b670-460dcecb6c68","Type":"ContainerStarted","Data":"bfbd7f055f41690d2395af1912ddb98be99cfbf2d8fdf33ff26d4d7653e8e8fa"} Mar 18 00:34:02 crc kubenswrapper[4983]: I0318 00:34:02.492429 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29563234-pzlvr" podStartSLOduration=1.192011728 podStartE2EDuration="2.492411667s" podCreationTimestamp="2026-03-18 00:34:00 +0000 UTC" firstStartedPulling="2026-03-18 00:34:00.723255906 +0000 UTC m=+1558.620630516" lastFinishedPulling="2026-03-18 00:34:02.023655845 +0000 UTC m=+1559.921030455" observedRunningTime="2026-03-18 00:34:02.492227642 +0000 UTC m=+1560.389602262" watchObservedRunningTime="2026-03-18 00:34:02.492411667 +0000 UTC m=+1560.389786277" Mar 18 00:34:03 crc kubenswrapper[4983]: I0318 00:34:03.492393 4983 generic.go:334] "Generic (PLEG): container finished" podID="5f02470c-1df4-4c79-b670-460dcecb6c68" containerID="bfbd7f055f41690d2395af1912ddb98be99cfbf2d8fdf33ff26d4d7653e8e8fa" exitCode=0 Mar 18 00:34:03 crc kubenswrapper[4983]: I0318 00:34:03.492442 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563234-pzlvr" event={"ID":"5f02470c-1df4-4c79-b670-460dcecb6c68","Type":"ContainerDied","Data":"bfbd7f055f41690d2395af1912ddb98be99cfbf2d8fdf33ff26d4d7653e8e8fa"} Mar 18 00:34:08 crc kubenswrapper[4983]: I0318 00:34:08.845792 4983 scope.go:117] "RemoveContainer" containerID="bd0392b28bda01f0b79a3679f23a700610d020a81d8cd8d02978d3716ea179b7" Mar 18 00:34:09 crc kubenswrapper[4983]: I0318 00:34:09.387556 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563234-pzlvr" Mar 18 00:34:09 crc kubenswrapper[4983]: I0318 00:34:09.543444 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563234-pzlvr" event={"ID":"5f02470c-1df4-4c79-b670-460dcecb6c68","Type":"ContainerDied","Data":"09c6d163e139400f12ee181ab5c88d790fc8718b23099b41b6fcff517e80629f"} Mar 18 00:34:09 crc kubenswrapper[4983]: I0318 00:34:09.543479 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09c6d163e139400f12ee181ab5c88d790fc8718b23099b41b6fcff517e80629f" Mar 18 00:34:09 crc kubenswrapper[4983]: I0318 00:34:09.543497 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563234-pzlvr" Mar 18 00:34:09 crc kubenswrapper[4983]: I0318 00:34:09.553626 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d7zf\" (UniqueName: \"kubernetes.io/projected/5f02470c-1df4-4c79-b670-460dcecb6c68-kube-api-access-8d7zf\") pod \"5f02470c-1df4-4c79-b670-460dcecb6c68\" (UID: \"5f02470c-1df4-4c79-b670-460dcecb6c68\") " Mar 18 00:34:09 crc kubenswrapper[4983]: I0318 00:34:09.573625 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f02470c-1df4-4c79-b670-460dcecb6c68-kube-api-access-8d7zf" (OuterVolumeSpecName: "kube-api-access-8d7zf") pod "5f02470c-1df4-4c79-b670-460dcecb6c68" (UID: "5f02470c-1df4-4c79-b670-460dcecb6c68"). InnerVolumeSpecName "kube-api-access-8d7zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:34:09 crc kubenswrapper[4983]: I0318 00:34:09.655308 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d7zf\" (UniqueName: \"kubernetes.io/projected/5f02470c-1df4-4c79-b670-460dcecb6c68-kube-api-access-8d7zf\") on node \"crc\" DevicePath \"\"" Mar 18 00:34:09 crc kubenswrapper[4983]: I0318 00:34:09.845244 4983 scope.go:117] "RemoveContainer" containerID="d07e637f0508f4034d3a16a86642f2c142b78eed7963870c276506e93bc14bd5" Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.471614 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563228-bj5lk"] Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.476801 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563228-bj5lk"] Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.566164 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g" event={"ID":"fcaaaeb3-eae9-44ee-8fc9-348df5c997a9","Type":"ContainerStarted","Data":"d4ae2da23d27be029f7c003a232023bc3c2cd20dd6bc1e09041f232099c69635"} Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.574736 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq" event={"ID":"65a13175-2c14-45b7-92f5-4677c93be37f","Type":"ContainerStarted","Data":"b0b31e4e3745f5e64b65d426c482351d12f744ac5f81d9c9a9fed4c5459a8fe3"} Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.578895 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"c0983273-84f3-4f26-bef5-e738d038c3d6","Type":"ContainerStarted","Data":"dbbebd123f00cb1261f8320d1d06ed113b4532accd363deb60d243bf848c9aac"} Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.629359 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/qdr-test" podStartSLOduration=1.863865964 podStartE2EDuration="11.629341794s" podCreationTimestamp="2026-03-18 00:33:59 +0000 UTC" firstStartedPulling="2026-03-18 00:34:00.036853878 +0000 UTC m=+1557.934228528" lastFinishedPulling="2026-03-18 00:34:09.802329748 +0000 UTC m=+1567.699704358" observedRunningTime="2026-03-18 00:34:10.625030332 +0000 UTC m=+1568.522404942" watchObservedRunningTime="2026-03-18 00:34:10.629341794 +0000 UTC m=+1568.526716404" Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.853180 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84ef092c-1aff-4381-8f9d-ec59958c1c1f" path="/var/lib/kubelet/pods/84ef092c-1aff-4381-8f9d-ec59958c1c1f/volumes" Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.930476 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/stf-smoketest-smoke1-9hhvj"] Mar 18 00:34:10 crc kubenswrapper[4983]: E0318 00:34:10.930879 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f02470c-1df4-4c79-b670-460dcecb6c68" containerName="oc" Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.930901 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f02470c-1df4-4c79-b670-460dcecb6c68" containerName="oc" Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.931110 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f02470c-1df4-4c79-b670-460dcecb6c68" containerName="oc" Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.935816 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.938390 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-entrypoint-script" Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.939918 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-9hhvj"] Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.942117 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-entrypoint-script" Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.942225 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-config" Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.942230 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-healthcheck-log" Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.942311 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-publisher" Mar 18 00:34:10 crc kubenswrapper[4983]: I0318 00:34:10.942402 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-sensubility-config" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.083377 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-healthcheck-log\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.083455 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-sensubility-config\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.083477 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-ceilometer-publisher\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.083537 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-collectd-config\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.083568 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rbg6\" (UniqueName: \"kubernetes.io/projected/831732d4-4747-45b3-aeef-6e4f846aad69-kube-api-access-4rbg6\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.083640 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.083679 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.185415 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-sensubility-config\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.185519 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-ceilometer-publisher\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.185610 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-collectd-config\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.185670 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rbg6\" (UniqueName: \"kubernetes.io/projected/831732d4-4747-45b3-aeef-6e4f846aad69-kube-api-access-4rbg6\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.185755 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.185795 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.185852 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-healthcheck-log\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.186557 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-ceilometer-publisher\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.187374 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-collectd-config\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.187402 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.187739 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-healthcheck-log\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.187998 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-sensubility-config\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.188132 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.214139 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/curl"] Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.215529 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.216901 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rbg6\" (UniqueName: \"kubernetes.io/projected/831732d4-4747-45b3-aeef-6e4f846aad69-kube-api-access-4rbg6\") pod \"stf-smoketest-smoke1-9hhvj\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.239835 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.250091 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.388397 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr97g\" (UniqueName: \"kubernetes.io/projected/032575be-ab42-46a7-92c7-d1160e9a4ae5-kube-api-access-hr97g\") pod \"curl\" (UID: \"032575be-ab42-46a7-92c7-d1160e9a4ae5\") " pod="service-telemetry/curl" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.489724 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr97g\" (UniqueName: \"kubernetes.io/projected/032575be-ab42-46a7-92c7-d1160e9a4ae5-kube-api-access-hr97g\") pod \"curl\" (UID: \"032575be-ab42-46a7-92c7-d1160e9a4ae5\") " pod="service-telemetry/curl" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.513501 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr97g\" (UniqueName: \"kubernetes.io/projected/032575be-ab42-46a7-92c7-d1160e9a4ae5-kube-api-access-hr97g\") pod \"curl\" (UID: \"032575be-ab42-46a7-92c7-d1160e9a4ae5\") " pod="service-telemetry/curl" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.564916 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.686118 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-9hhvj"] Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.845343 4983 scope.go:117] "RemoveContainer" containerID="ee6750b5c3ce747e6604f74fb7b704f65681d582eab3c6429e39e89c6be1b546" Mar 18 00:34:11 crc kubenswrapper[4983]: I0318 00:34:11.859395 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Mar 18 00:34:11 crc kubenswrapper[4983]: W0318 00:34:11.895029 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod032575be_ab42_46a7_92c7_d1160e9a4ae5.slice/crio-dc0fc5695ceecf59e85b73589a87d1f12599d0ba40bb0f2e6deb515717f5db67 WatchSource:0}: Error finding container dc0fc5695ceecf59e85b73589a87d1f12599d0ba40bb0f2e6deb515717f5db67: Status 404 returned error can't find the container with id dc0fc5695ceecf59e85b73589a87d1f12599d0ba40bb0f2e6deb515717f5db67 Mar 18 00:34:12 crc kubenswrapper[4983]: I0318 00:34:12.627937 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-9hhvj" event={"ID":"831732d4-4747-45b3-aeef-6e4f846aad69","Type":"ContainerStarted","Data":"0686090128ce4746269e718a4ba54e96623182eac1b8d24d4f1d0775e9c290f7"} Mar 18 00:34:12 crc kubenswrapper[4983]: I0318 00:34:12.636878 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7" event={"ID":"937b4985-05d5-48c0-b408-3164619e74ca","Type":"ContainerStarted","Data":"02386c886875f7aca13a4f575acfe749f7665b99dc7a742ab598e837c5dad980"} Mar 18 00:34:12 crc kubenswrapper[4983]: I0318 00:34:12.643521 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"032575be-ab42-46a7-92c7-d1160e9a4ae5","Type":"ContainerStarted","Data":"dc0fc5695ceecf59e85b73589a87d1f12599d0ba40bb0f2e6deb515717f5db67"} Mar 18 00:34:12 crc kubenswrapper[4983]: I0318 00:34:12.849474 4983 scope.go:117] "RemoveContainer" containerID="64f6fb2c5a225889892516fb2259bd64047b795f6a443c063da26ff7cc7e305e" Mar 18 00:34:12 crc kubenswrapper[4983]: I0318 00:34:12.849887 4983 scope.go:117] "RemoveContainer" containerID="7be6a908d8aea62e2c9ee64bf8187c8c256f998e60416d9e03b4b2b349ca9e6b" Mar 18 00:34:13 crc kubenswrapper[4983]: I0318 00:34:13.259020 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:34:13 crc kubenswrapper[4983]: I0318 00:34:13.259387 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:34:13 crc kubenswrapper[4983]: I0318 00:34:13.259429 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:34:13 crc kubenswrapper[4983]: I0318 00:34:13.260085 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9"} pod="openshift-machine-config-operator/machine-config-daemon-jczfs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 00:34:13 crc kubenswrapper[4983]: I0318 00:34:13.260142 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" containerID="cri-o://fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" gracePeriod=600 Mar 18 00:34:13 crc kubenswrapper[4983]: E0318 00:34:13.459710 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:34:13 crc kubenswrapper[4983]: I0318 00:34:13.658779 4983 generic.go:334] "Generic (PLEG): container finished" podID="482e2afe-48be-4ea2-8bda-52849c901b52" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" exitCode=0 Mar 18 00:34:13 crc kubenswrapper[4983]: I0318 00:34:13.658859 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerDied","Data":"fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9"} Mar 18 00:34:13 crc kubenswrapper[4983]: I0318 00:34:13.658895 4983 scope.go:117] "RemoveContainer" containerID="daa639b5561e05c04168c0703cff853339f7b4e1eea5d3b72dde7a611b097532" Mar 18 00:34:13 crc kubenswrapper[4983]: I0318 00:34:13.659380 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:34:13 crc kubenswrapper[4983]: E0318 00:34:13.659571 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:34:13 crc kubenswrapper[4983]: I0318 00:34:13.669407 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp" event={"ID":"3753b92e-109f-4875-a1fe-70ea96a52a29","Type":"ContainerStarted","Data":"45d022e53ceb9572520fd0ca3665200e8fbcf3f1967d284702140267cef648fa"} Mar 18 00:34:13 crc kubenswrapper[4983]: I0318 00:34:13.684940 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2" event={"ID":"2ba80112-3269-42af-a92b-86876fe20c01","Type":"ContainerStarted","Data":"19d5faf5ff01e2ae06fb4fe7355c70b702a330b74c6772713bd9caff2e324978"} Mar 18 00:34:14 crc kubenswrapper[4983]: I0318 00:34:14.697080 4983 generic.go:334] "Generic (PLEG): container finished" podID="032575be-ab42-46a7-92c7-d1160e9a4ae5" containerID="7b70bfac2fde065eedf13e937cfbeafa9531f0db57b5b95169e6716f41d321cb" exitCode=0 Mar 18 00:34:14 crc kubenswrapper[4983]: I0318 00:34:14.697127 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"032575be-ab42-46a7-92c7-d1160e9a4ae5","Type":"ContainerDied","Data":"7b70bfac2fde065eedf13e937cfbeafa9531f0db57b5b95169e6716f41d321cb"} Mar 18 00:34:21 crc kubenswrapper[4983]: I0318 00:34:21.840288 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 18 00:34:21 crc kubenswrapper[4983]: I0318 00:34:21.962355 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr97g\" (UniqueName: \"kubernetes.io/projected/032575be-ab42-46a7-92c7-d1160e9a4ae5-kube-api-access-hr97g\") pod \"032575be-ab42-46a7-92c7-d1160e9a4ae5\" (UID: \"032575be-ab42-46a7-92c7-d1160e9a4ae5\") " Mar 18 00:34:21 crc kubenswrapper[4983]: I0318 00:34:21.973085 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/032575be-ab42-46a7-92c7-d1160e9a4ae5-kube-api-access-hr97g" (OuterVolumeSpecName: "kube-api-access-hr97g") pod "032575be-ab42-46a7-92c7-d1160e9a4ae5" (UID: "032575be-ab42-46a7-92c7-d1160e9a4ae5"). InnerVolumeSpecName "kube-api-access-hr97g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:34:22 crc kubenswrapper[4983]: I0318 00:34:22.037063 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_curl_032575be-ab42-46a7-92c7-d1160e9a4ae5/curl/0.log" Mar 18 00:34:22 crc kubenswrapper[4983]: I0318 00:34:22.064679 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr97g\" (UniqueName: \"kubernetes.io/projected/032575be-ab42-46a7-92c7-d1160e9a4ae5-kube-api-access-hr97g\") on node \"crc\" DevicePath \"\"" Mar 18 00:34:22 crc kubenswrapper[4983]: I0318 00:34:22.294219 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-mqksz_c025a69f-f942-4648-bb53-972903a8e14d/prometheus-webhook-snmp/0.log" Mar 18 00:34:22 crc kubenswrapper[4983]: I0318 00:34:22.774938 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 18 00:34:22 crc kubenswrapper[4983]: I0318 00:34:22.774836 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"032575be-ab42-46a7-92c7-d1160e9a4ae5","Type":"ContainerDied","Data":"dc0fc5695ceecf59e85b73589a87d1f12599d0ba40bb0f2e6deb515717f5db67"} Mar 18 00:34:22 crc kubenswrapper[4983]: I0318 00:34:22.782938 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc0fc5695ceecf59e85b73589a87d1f12599d0ba40bb0f2e6deb515717f5db67" Mar 18 00:34:23 crc kubenswrapper[4983]: I0318 00:34:23.781437 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-9hhvj" event={"ID":"831732d4-4747-45b3-aeef-6e4f846aad69","Type":"ContainerStarted","Data":"afe9ceba8a39d1c6911367925a6981b4a5bc18ce26d8afdc22f4efeeda78c5ae"} Mar 18 00:34:24 crc kubenswrapper[4983]: I0318 00:34:24.844956 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:34:24 crc kubenswrapper[4983]: E0318 00:34:24.845480 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:34:29 crc kubenswrapper[4983]: I0318 00:34:29.829188 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-9hhvj" event={"ID":"831732d4-4747-45b3-aeef-6e4f846aad69","Type":"ContainerStarted","Data":"e5f6a07bf293ed1b1888f728345e259e42d4c2f1a5dd7c5e75c9fbb01aa959c0"} Mar 18 00:34:29 crc kubenswrapper[4983]: I0318 00:34:29.861818 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/stf-smoketest-smoke1-9hhvj" podStartSLOduration=2.737178176 podStartE2EDuration="19.861796462s" podCreationTimestamp="2026-03-18 00:34:10 +0000 UTC" firstStartedPulling="2026-03-18 00:34:11.706052286 +0000 UTC m=+1569.603426886" lastFinishedPulling="2026-03-18 00:34:28.830670562 +0000 UTC m=+1586.728045172" observedRunningTime="2026-03-18 00:34:29.855974568 +0000 UTC m=+1587.753349208" watchObservedRunningTime="2026-03-18 00:34:29.861796462 +0000 UTC m=+1587.759171072" Mar 18 00:34:37 crc kubenswrapper[4983]: I0318 00:34:37.844833 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:34:37 crc kubenswrapper[4983]: E0318 00:34:37.847308 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:34:50 crc kubenswrapper[4983]: I0318 00:34:50.844676 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:34:50 crc kubenswrapper[4983]: E0318 00:34:50.846073 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:34:52 crc kubenswrapper[4983]: I0318 00:34:52.425229 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-mqksz_c025a69f-f942-4648-bb53-972903a8e14d/prometheus-webhook-snmp/0.log" Mar 18 00:34:58 crc kubenswrapper[4983]: I0318 00:34:58.097886 4983 generic.go:334] "Generic (PLEG): container finished" podID="831732d4-4747-45b3-aeef-6e4f846aad69" containerID="afe9ceba8a39d1c6911367925a6981b4a5bc18ce26d8afdc22f4efeeda78c5ae" exitCode=0 Mar 18 00:34:58 crc kubenswrapper[4983]: I0318 00:34:58.098107 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-9hhvj" event={"ID":"831732d4-4747-45b3-aeef-6e4f846aad69","Type":"ContainerDied","Data":"afe9ceba8a39d1c6911367925a6981b4a5bc18ce26d8afdc22f4efeeda78c5ae"} Mar 18 00:34:58 crc kubenswrapper[4983]: I0318 00:34:58.099314 4983 scope.go:117] "RemoveContainer" containerID="afe9ceba8a39d1c6911367925a6981b4a5bc18ce26d8afdc22f4efeeda78c5ae" Mar 18 00:35:01 crc kubenswrapper[4983]: I0318 00:35:01.488631 4983 generic.go:334] "Generic (PLEG): container finished" podID="831732d4-4747-45b3-aeef-6e4f846aad69" containerID="e5f6a07bf293ed1b1888f728345e259e42d4c2f1a5dd7c5e75c9fbb01aa959c0" exitCode=0 Mar 18 00:35:01 crc kubenswrapper[4983]: I0318 00:35:01.488750 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-9hhvj" event={"ID":"831732d4-4747-45b3-aeef-6e4f846aad69","Type":"ContainerDied","Data":"e5f6a07bf293ed1b1888f728345e259e42d4c2f1a5dd7c5e75c9fbb01aa959c0"} Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.841420 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.844036 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:35:02 crc kubenswrapper[4983]: E0318 00:35:02.844261 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.885597 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-ceilometer-publisher\") pod \"831732d4-4747-45b3-aeef-6e4f846aad69\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.886233 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-ceilometer-entrypoint-script\") pod \"831732d4-4747-45b3-aeef-6e4f846aad69\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.886258 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-sensubility-config\") pod \"831732d4-4747-45b3-aeef-6e4f846aad69\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.886315 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-collectd-config\") pod \"831732d4-4747-45b3-aeef-6e4f846aad69\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.886347 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-collectd-entrypoint-script\") pod \"831732d4-4747-45b3-aeef-6e4f846aad69\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.886446 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rbg6\" (UniqueName: \"kubernetes.io/projected/831732d4-4747-45b3-aeef-6e4f846aad69-kube-api-access-4rbg6\") pod \"831732d4-4747-45b3-aeef-6e4f846aad69\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.886492 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-healthcheck-log\") pod \"831732d4-4747-45b3-aeef-6e4f846aad69\" (UID: \"831732d4-4747-45b3-aeef-6e4f846aad69\") " Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.893217 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/831732d4-4747-45b3-aeef-6e4f846aad69-kube-api-access-4rbg6" (OuterVolumeSpecName: "kube-api-access-4rbg6") pod "831732d4-4747-45b3-aeef-6e4f846aad69" (UID: "831732d4-4747-45b3-aeef-6e4f846aad69"). InnerVolumeSpecName "kube-api-access-4rbg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.906923 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-collectd-config" (OuterVolumeSpecName: "collectd-config") pod "831732d4-4747-45b3-aeef-6e4f846aad69" (UID: "831732d4-4747-45b3-aeef-6e4f846aad69"). InnerVolumeSpecName "collectd-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.908880 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-healthcheck-log" (OuterVolumeSpecName: "healthcheck-log") pod "831732d4-4747-45b3-aeef-6e4f846aad69" (UID: "831732d4-4747-45b3-aeef-6e4f846aad69"). InnerVolumeSpecName "healthcheck-log". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.909488 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-collectd-entrypoint-script" (OuterVolumeSpecName: "collectd-entrypoint-script") pod "831732d4-4747-45b3-aeef-6e4f846aad69" (UID: "831732d4-4747-45b3-aeef-6e4f846aad69"). InnerVolumeSpecName "collectd-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.913826 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-sensubility-config" (OuterVolumeSpecName: "sensubility-config") pod "831732d4-4747-45b3-aeef-6e4f846aad69" (UID: "831732d4-4747-45b3-aeef-6e4f846aad69"). InnerVolumeSpecName "sensubility-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.924136 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-ceilometer-publisher" (OuterVolumeSpecName: "ceilometer-publisher") pod "831732d4-4747-45b3-aeef-6e4f846aad69" (UID: "831732d4-4747-45b3-aeef-6e4f846aad69"). InnerVolumeSpecName "ceilometer-publisher". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.925040 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-ceilometer-entrypoint-script" (OuterVolumeSpecName: "ceilometer-entrypoint-script") pod "831732d4-4747-45b3-aeef-6e4f846aad69" (UID: "831732d4-4747-45b3-aeef-6e4f846aad69"). InnerVolumeSpecName "ceilometer-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.988765 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rbg6\" (UniqueName: \"kubernetes.io/projected/831732d4-4747-45b3-aeef-6e4f846aad69-kube-api-access-4rbg6\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.988819 4983 reconciler_common.go:293] "Volume detached for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-healthcheck-log\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.988841 4983 reconciler_common.go:293] "Volume detached for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-ceilometer-publisher\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.988859 4983 reconciler_common.go:293] "Volume detached for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-ceilometer-entrypoint-script\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.988879 4983 reconciler_common.go:293] "Volume detached for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-sensubility-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.988899 4983 reconciler_common.go:293] "Volume detached for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-collectd-config\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:02 crc kubenswrapper[4983]: I0318 00:35:02.988916 4983 reconciler_common.go:293] "Volume detached for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/831732d4-4747-45b3-aeef-6e4f846aad69-collectd-entrypoint-script\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:03 crc kubenswrapper[4983]: I0318 00:35:03.509417 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-9hhvj" event={"ID":"831732d4-4747-45b3-aeef-6e4f846aad69","Type":"ContainerDied","Data":"0686090128ce4746269e718a4ba54e96623182eac1b8d24d4f1d0775e9c290f7"} Mar 18 00:35:03 crc kubenswrapper[4983]: I0318 00:35:03.509487 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0686090128ce4746269e718a4ba54e96623182eac1b8d24d4f1d0775e9c290f7" Mar 18 00:35:03 crc kubenswrapper[4983]: I0318 00:35:03.509495 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-9hhvj" Mar 18 00:35:04 crc kubenswrapper[4983]: I0318 00:35:04.862620 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-9hhvj_831732d4-4747-45b3-aeef-6e4f846aad69/smoketest-collectd/0.log" Mar 18 00:35:05 crc kubenswrapper[4983]: I0318 00:35:05.092964 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-9hhvj_831732d4-4747-45b3-aeef-6e4f846aad69/smoketest-ceilometer/0.log" Mar 18 00:35:05 crc kubenswrapper[4983]: I0318 00:35:05.352191 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-interconnect-68864d46cb-jn22d_194c9fd4-aedf-4d03-a82d-077903e33e44/default-interconnect/0.log" Mar 18 00:35:05 crc kubenswrapper[4983]: I0318 00:35:05.585882 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq_65a13175-2c14-45b7-92f5-4677c93be37f/bridge/2.log" Mar 18 00:35:05 crc kubenswrapper[4983]: I0318 00:35:05.856894 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-5t9zq_65a13175-2c14-45b7-92f5-4677c93be37f/sg-core/0.log" Mar 18 00:35:06 crc kubenswrapper[4983]: I0318 00:35:06.104620 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2_2ba80112-3269-42af-a92b-86876fe20c01/bridge/2.log" Mar 18 00:35:06 crc kubenswrapper[4983]: I0318 00:35:06.321244 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-6dc8b66747-pqdj2_2ba80112-3269-42af-a92b-86876fe20c01/sg-core/0.log" Mar 18 00:35:06 crc kubenswrapper[4983]: I0318 00:35:06.532039 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7_937b4985-05d5-48c0-b408-3164619e74ca/bridge/2.log" Mar 18 00:35:06 crc kubenswrapper[4983]: I0318 00:35:06.757236 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-kh6l7_937b4985-05d5-48c0-b408-3164619e74ca/sg-core/0.log" Mar 18 00:35:07 crc kubenswrapper[4983]: I0318 00:35:07.017368 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g_fcaaaeb3-eae9-44ee-8fc9-348df5c997a9/bridge/2.log" Mar 18 00:35:07 crc kubenswrapper[4983]: I0318 00:35:07.241466 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-f55dd7b74-b4h8g_fcaaaeb3-eae9-44ee-8fc9-348df5c997a9/sg-core/0.log" Mar 18 00:35:07 crc kubenswrapper[4983]: I0318 00:35:07.467171 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp_3753b92e-109f-4875-a1fe-70ea96a52a29/bridge/2.log" Mar 18 00:35:07 crc kubenswrapper[4983]: I0318 00:35:07.719811 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-vg8xp_3753b92e-109f-4875-a1fe-70ea96a52a29/sg-core/0.log" Mar 18 00:35:10 crc kubenswrapper[4983]: I0318 00:35:10.447367 4983 scope.go:117] "RemoveContainer" containerID="fd2afaab66d567abc373f78c2e812ec98a921ec7fb763c221086deb26161cfe1" Mar 18 00:35:11 crc kubenswrapper[4983]: I0318 00:35:11.790984 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-79ff6f684b-49b74_71ccd954-e60f-4392-9ce6-3672ff3cadce/operator/0.log" Mar 18 00:35:12 crc kubenswrapper[4983]: I0318 00:35:12.087884 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_ec767daa-3f6a-4501-b63a-dd8949ef6199/prometheus/0.log" Mar 18 00:35:12 crc kubenswrapper[4983]: I0318 00:35:12.376469 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_682dc305-300d-4702-a640-b422a2185661/elasticsearch/0.log" Mar 18 00:35:12 crc kubenswrapper[4983]: I0318 00:35:12.595546 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-mqksz_c025a69f-f942-4648-bb53-972903a8e14d/prometheus-webhook-snmp/0.log" Mar 18 00:35:12 crc kubenswrapper[4983]: I0318 00:35:12.867699 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_6c1ecce7-261c-4f82-863c-641ef76200ae/alertmanager/0.log" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.560371 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lfz8b"] Mar 18 00:35:14 crc kubenswrapper[4983]: E0318 00:35:14.561219 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="831732d4-4747-45b3-aeef-6e4f846aad69" containerName="smoketest-collectd" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.561242 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="831732d4-4747-45b3-aeef-6e4f846aad69" containerName="smoketest-collectd" Mar 18 00:35:14 crc kubenswrapper[4983]: E0318 00:35:14.561270 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="831732d4-4747-45b3-aeef-6e4f846aad69" containerName="smoketest-ceilometer" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.561282 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="831732d4-4747-45b3-aeef-6e4f846aad69" containerName="smoketest-ceilometer" Mar 18 00:35:14 crc kubenswrapper[4983]: E0318 00:35:14.561306 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032575be-ab42-46a7-92c7-d1160e9a4ae5" containerName="curl" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.561319 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="032575be-ab42-46a7-92c7-d1160e9a4ae5" containerName="curl" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.561559 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="831732d4-4747-45b3-aeef-6e4f846aad69" containerName="smoketest-collectd" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.561578 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="831732d4-4747-45b3-aeef-6e4f846aad69" containerName="smoketest-ceilometer" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.561595 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="032575be-ab42-46a7-92c7-d1160e9a4ae5" containerName="curl" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.567692 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.579118 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lfz8b"] Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.694937 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nxzl\" (UniqueName: \"kubernetes.io/projected/2437ad60-b755-4cc9-ad65-084d8b4ca03e-kube-api-access-2nxzl\") pod \"redhat-operators-lfz8b\" (UID: \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\") " pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.695148 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2437ad60-b755-4cc9-ad65-084d8b4ca03e-utilities\") pod \"redhat-operators-lfz8b\" (UID: \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\") " pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.695202 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2437ad60-b755-4cc9-ad65-084d8b4ca03e-catalog-content\") pod \"redhat-operators-lfz8b\" (UID: \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\") " pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.797062 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2437ad60-b755-4cc9-ad65-084d8b4ca03e-utilities\") pod \"redhat-operators-lfz8b\" (UID: \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\") " pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.797150 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2437ad60-b755-4cc9-ad65-084d8b4ca03e-catalog-content\") pod \"redhat-operators-lfz8b\" (UID: \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\") " pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.797236 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nxzl\" (UniqueName: \"kubernetes.io/projected/2437ad60-b755-4cc9-ad65-084d8b4ca03e-kube-api-access-2nxzl\") pod \"redhat-operators-lfz8b\" (UID: \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\") " pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.797677 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2437ad60-b755-4cc9-ad65-084d8b4ca03e-utilities\") pod \"redhat-operators-lfz8b\" (UID: \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\") " pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.797695 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2437ad60-b755-4cc9-ad65-084d8b4ca03e-catalog-content\") pod \"redhat-operators-lfz8b\" (UID: \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\") " pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.819148 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nxzl\" (UniqueName: \"kubernetes.io/projected/2437ad60-b755-4cc9-ad65-084d8b4ca03e-kube-api-access-2nxzl\") pod \"redhat-operators-lfz8b\" (UID: \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\") " pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.845771 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:35:14 crc kubenswrapper[4983]: E0318 00:35:14.846577 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:35:14 crc kubenswrapper[4983]: I0318 00:35:14.889958 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:15 crc kubenswrapper[4983]: I0318 00:35:15.333037 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lfz8b"] Mar 18 00:35:15 crc kubenswrapper[4983]: I0318 00:35:15.633187 4983 generic.go:334] "Generic (PLEG): container finished" podID="2437ad60-b755-4cc9-ad65-084d8b4ca03e" containerID="f7bb2c2226d74d71536f9100c7f562c13d7876e0020511233163d67ea55f2f39" exitCode=0 Mar 18 00:35:15 crc kubenswrapper[4983]: I0318 00:35:15.633257 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfz8b" event={"ID":"2437ad60-b755-4cc9-ad65-084d8b4ca03e","Type":"ContainerDied","Data":"f7bb2c2226d74d71536f9100c7f562c13d7876e0020511233163d67ea55f2f39"} Mar 18 00:35:15 crc kubenswrapper[4983]: I0318 00:35:15.633641 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfz8b" event={"ID":"2437ad60-b755-4cc9-ad65-084d8b4ca03e","Type":"ContainerStarted","Data":"2f4828f58482c373836d3f3da3f04f59308b714ca6ec25c2ca17a731686ad637"} Mar 18 00:35:15 crc kubenswrapper[4983]: I0318 00:35:15.635977 4983 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 00:35:17 crc kubenswrapper[4983]: I0318 00:35:17.654298 4983 generic.go:334] "Generic (PLEG): container finished" podID="2437ad60-b755-4cc9-ad65-084d8b4ca03e" containerID="1376843a0a09d58d678b53577403712502643d21cc36f5a25b475f6f0abd97db" exitCode=0 Mar 18 00:35:17 crc kubenswrapper[4983]: I0318 00:35:17.654381 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfz8b" event={"ID":"2437ad60-b755-4cc9-ad65-084d8b4ca03e","Type":"ContainerDied","Data":"1376843a0a09d58d678b53577403712502643d21cc36f5a25b475f6f0abd97db"} Mar 18 00:35:18 crc kubenswrapper[4983]: I0318 00:35:18.671314 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfz8b" event={"ID":"2437ad60-b755-4cc9-ad65-084d8b4ca03e","Type":"ContainerStarted","Data":"7b3fd9b32e3354a48acbc973efe1276b89c787168b7b97543d79ede73fbe6c25"} Mar 18 00:35:18 crc kubenswrapper[4983]: I0318 00:35:18.691470 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lfz8b" podStartSLOduration=2.206040298 podStartE2EDuration="4.691449058s" podCreationTimestamp="2026-03-18 00:35:14 +0000 UTC" firstStartedPulling="2026-03-18 00:35:15.635708346 +0000 UTC m=+1633.533082956" lastFinishedPulling="2026-03-18 00:35:18.121117106 +0000 UTC m=+1636.018491716" observedRunningTime="2026-03-18 00:35:18.686775216 +0000 UTC m=+1636.584149826" watchObservedRunningTime="2026-03-18 00:35:18.691449058 +0000 UTC m=+1636.588823668" Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.121496 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wkdxj"] Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.123077 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.141679 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wkdxj"] Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.165980 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxknc\" (UniqueName: \"kubernetes.io/projected/35efb54b-6131-47cd-b092-28187491df7f-kube-api-access-dxknc\") pod \"certified-operators-wkdxj\" (UID: \"35efb54b-6131-47cd-b092-28187491df7f\") " pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.166048 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35efb54b-6131-47cd-b092-28187491df7f-catalog-content\") pod \"certified-operators-wkdxj\" (UID: \"35efb54b-6131-47cd-b092-28187491df7f\") " pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.166122 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35efb54b-6131-47cd-b092-28187491df7f-utilities\") pod \"certified-operators-wkdxj\" (UID: \"35efb54b-6131-47cd-b092-28187491df7f\") " pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.267842 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35efb54b-6131-47cd-b092-28187491df7f-catalog-content\") pod \"certified-operators-wkdxj\" (UID: \"35efb54b-6131-47cd-b092-28187491df7f\") " pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.267979 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35efb54b-6131-47cd-b092-28187491df7f-utilities\") pod \"certified-operators-wkdxj\" (UID: \"35efb54b-6131-47cd-b092-28187491df7f\") " pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.268067 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxknc\" (UniqueName: \"kubernetes.io/projected/35efb54b-6131-47cd-b092-28187491df7f-kube-api-access-dxknc\") pod \"certified-operators-wkdxj\" (UID: \"35efb54b-6131-47cd-b092-28187491df7f\") " pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.268492 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35efb54b-6131-47cd-b092-28187491df7f-catalog-content\") pod \"certified-operators-wkdxj\" (UID: \"35efb54b-6131-47cd-b092-28187491df7f\") " pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.268814 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35efb54b-6131-47cd-b092-28187491df7f-utilities\") pod \"certified-operators-wkdxj\" (UID: \"35efb54b-6131-47cd-b092-28187491df7f\") " pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.288191 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxknc\" (UniqueName: \"kubernetes.io/projected/35efb54b-6131-47cd-b092-28187491df7f-kube-api-access-dxknc\") pod \"certified-operators-wkdxj\" (UID: \"35efb54b-6131-47cd-b092-28187491df7f\") " pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.441117 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:19 crc kubenswrapper[4983]: I0318 00:35:19.709337 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wkdxj"] Mar 18 00:35:20 crc kubenswrapper[4983]: I0318 00:35:20.695757 4983 generic.go:334] "Generic (PLEG): container finished" podID="35efb54b-6131-47cd-b092-28187491df7f" containerID="7230bc0f9e808b6be448dd8cfa4c4203d28240580797cc32851f9de2ebbfff95" exitCode=0 Mar 18 00:35:20 crc kubenswrapper[4983]: I0318 00:35:20.696129 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wkdxj" event={"ID":"35efb54b-6131-47cd-b092-28187491df7f","Type":"ContainerDied","Data":"7230bc0f9e808b6be448dd8cfa4c4203d28240580797cc32851f9de2ebbfff95"} Mar 18 00:35:20 crc kubenswrapper[4983]: I0318 00:35:20.696166 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wkdxj" event={"ID":"35efb54b-6131-47cd-b092-28187491df7f","Type":"ContainerStarted","Data":"99fff2aa6b2bb90b6085cc0c8ea4ee280d9d9cb2610d392ec86d7570c9d25c74"} Mar 18 00:35:21 crc kubenswrapper[4983]: I0318 00:35:21.704433 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wkdxj" event={"ID":"35efb54b-6131-47cd-b092-28187491df7f","Type":"ContainerStarted","Data":"edf0397b7b8e3127ca17cfd8f9a53ae29dbd5aa403b95ff6fbf64049fe7163f1"} Mar 18 00:35:22 crc kubenswrapper[4983]: I0318 00:35:22.719276 4983 generic.go:334] "Generic (PLEG): container finished" podID="35efb54b-6131-47cd-b092-28187491df7f" containerID="edf0397b7b8e3127ca17cfd8f9a53ae29dbd5aa403b95ff6fbf64049fe7163f1" exitCode=0 Mar 18 00:35:22 crc kubenswrapper[4983]: I0318 00:35:22.719347 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wkdxj" event={"ID":"35efb54b-6131-47cd-b092-28187491df7f","Type":"ContainerDied","Data":"edf0397b7b8e3127ca17cfd8f9a53ae29dbd5aa403b95ff6fbf64049fe7163f1"} Mar 18 00:35:23 crc kubenswrapper[4983]: I0318 00:35:23.731420 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wkdxj" event={"ID":"35efb54b-6131-47cd-b092-28187491df7f","Type":"ContainerStarted","Data":"ac23ccd8047a2d2bd9675277e893feb400b1ffda959e081c6702b0aa7701e375"} Mar 18 00:35:23 crc kubenswrapper[4983]: I0318 00:35:23.758146 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wkdxj" podStartSLOduration=2.230144746 podStartE2EDuration="4.758126564s" podCreationTimestamp="2026-03-18 00:35:19 +0000 UTC" firstStartedPulling="2026-03-18 00:35:20.697736472 +0000 UTC m=+1638.595111092" lastFinishedPulling="2026-03-18 00:35:23.2257183 +0000 UTC m=+1641.123092910" observedRunningTime="2026-03-18 00:35:23.751087036 +0000 UTC m=+1641.648461656" watchObservedRunningTime="2026-03-18 00:35:23.758126564 +0000 UTC m=+1641.655501174" Mar 18 00:35:24 crc kubenswrapper[4983]: I0318 00:35:24.890381 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:24 crc kubenswrapper[4983]: I0318 00:35:24.891579 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:25 crc kubenswrapper[4983]: I0318 00:35:25.938466 4983 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lfz8b" podUID="2437ad60-b755-4cc9-ad65-084d8b4ca03e" containerName="registry-server" probeResult="failure" output=< Mar 18 00:35:25 crc kubenswrapper[4983]: timeout: failed to connect service ":50051" within 1s Mar 18 00:35:25 crc kubenswrapper[4983]: > Mar 18 00:35:26 crc kubenswrapper[4983]: I0318 00:35:26.387415 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-7f455c87c8-7n4mj_6baa739d-8869-4888-95b8-821fa42715b7/operator/0.log" Mar 18 00:35:26 crc kubenswrapper[4983]: I0318 00:35:26.844715 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:35:26 crc kubenswrapper[4983]: E0318 00:35:26.845163 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:35:29 crc kubenswrapper[4983]: I0318 00:35:29.442022 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:29 crc kubenswrapper[4983]: I0318 00:35:29.443206 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:29 crc kubenswrapper[4983]: I0318 00:35:29.496598 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:29 crc kubenswrapper[4983]: I0318 00:35:29.572435 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-79ff6f684b-49b74_71ccd954-e60f-4392-9ce6-3672ff3cadce/operator/0.log" Mar 18 00:35:29 crc kubenswrapper[4983]: I0318 00:35:29.800325 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_qdr-test_c0983273-84f3-4f26-bef5-e738d038c3d6/qdr/0.log" Mar 18 00:35:29 crc kubenswrapper[4983]: I0318 00:35:29.845779 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:29 crc kubenswrapper[4983]: I0318 00:35:29.907725 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wkdxj"] Mar 18 00:35:31 crc kubenswrapper[4983]: I0318 00:35:31.803332 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wkdxj" podUID="35efb54b-6131-47cd-b092-28187491df7f" containerName="registry-server" containerID="cri-o://ac23ccd8047a2d2bd9675277e893feb400b1ffda959e081c6702b0aa7701e375" gracePeriod=2 Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.197034 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.280745 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35efb54b-6131-47cd-b092-28187491df7f-catalog-content\") pod \"35efb54b-6131-47cd-b092-28187491df7f\" (UID: \"35efb54b-6131-47cd-b092-28187491df7f\") " Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.280862 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35efb54b-6131-47cd-b092-28187491df7f-utilities\") pod \"35efb54b-6131-47cd-b092-28187491df7f\" (UID: \"35efb54b-6131-47cd-b092-28187491df7f\") " Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.281078 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxknc\" (UniqueName: \"kubernetes.io/projected/35efb54b-6131-47cd-b092-28187491df7f-kube-api-access-dxknc\") pod \"35efb54b-6131-47cd-b092-28187491df7f\" (UID: \"35efb54b-6131-47cd-b092-28187491df7f\") " Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.281912 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35efb54b-6131-47cd-b092-28187491df7f-utilities" (OuterVolumeSpecName: "utilities") pod "35efb54b-6131-47cd-b092-28187491df7f" (UID: "35efb54b-6131-47cd-b092-28187491df7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.288884 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35efb54b-6131-47cd-b092-28187491df7f-kube-api-access-dxknc" (OuterVolumeSpecName: "kube-api-access-dxknc") pod "35efb54b-6131-47cd-b092-28187491df7f" (UID: "35efb54b-6131-47cd-b092-28187491df7f"). InnerVolumeSpecName "kube-api-access-dxknc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.335080 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35efb54b-6131-47cd-b092-28187491df7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35efb54b-6131-47cd-b092-28187491df7f" (UID: "35efb54b-6131-47cd-b092-28187491df7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.383070 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxknc\" (UniqueName: \"kubernetes.io/projected/35efb54b-6131-47cd-b092-28187491df7f-kube-api-access-dxknc\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.383395 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35efb54b-6131-47cd-b092-28187491df7f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.383405 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35efb54b-6131-47cd-b092-28187491df7f-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.813962 4983 generic.go:334] "Generic (PLEG): container finished" podID="35efb54b-6131-47cd-b092-28187491df7f" containerID="ac23ccd8047a2d2bd9675277e893feb400b1ffda959e081c6702b0aa7701e375" exitCode=0 Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.814032 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wkdxj" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.814032 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wkdxj" event={"ID":"35efb54b-6131-47cd-b092-28187491df7f","Type":"ContainerDied","Data":"ac23ccd8047a2d2bd9675277e893feb400b1ffda959e081c6702b0aa7701e375"} Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.814526 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wkdxj" event={"ID":"35efb54b-6131-47cd-b092-28187491df7f","Type":"ContainerDied","Data":"99fff2aa6b2bb90b6085cc0c8ea4ee280d9d9cb2610d392ec86d7570c9d25c74"} Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.814563 4983 scope.go:117] "RemoveContainer" containerID="ac23ccd8047a2d2bd9675277e893feb400b1ffda959e081c6702b0aa7701e375" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.848309 4983 scope.go:117] "RemoveContainer" containerID="edf0397b7b8e3127ca17cfd8f9a53ae29dbd5aa403b95ff6fbf64049fe7163f1" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.870665 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wkdxj"] Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.870707 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wkdxj"] Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.873390 4983 scope.go:117] "RemoveContainer" containerID="7230bc0f9e808b6be448dd8cfa4c4203d28240580797cc32851f9de2ebbfff95" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.893198 4983 scope.go:117] "RemoveContainer" containerID="ac23ccd8047a2d2bd9675277e893feb400b1ffda959e081c6702b0aa7701e375" Mar 18 00:35:32 crc kubenswrapper[4983]: E0318 00:35:32.894085 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac23ccd8047a2d2bd9675277e893feb400b1ffda959e081c6702b0aa7701e375\": container with ID starting with ac23ccd8047a2d2bd9675277e893feb400b1ffda959e081c6702b0aa7701e375 not found: ID does not exist" containerID="ac23ccd8047a2d2bd9675277e893feb400b1ffda959e081c6702b0aa7701e375" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.894140 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac23ccd8047a2d2bd9675277e893feb400b1ffda959e081c6702b0aa7701e375"} err="failed to get container status \"ac23ccd8047a2d2bd9675277e893feb400b1ffda959e081c6702b0aa7701e375\": rpc error: code = NotFound desc = could not find container \"ac23ccd8047a2d2bd9675277e893feb400b1ffda959e081c6702b0aa7701e375\": container with ID starting with ac23ccd8047a2d2bd9675277e893feb400b1ffda959e081c6702b0aa7701e375 not found: ID does not exist" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.894175 4983 scope.go:117] "RemoveContainer" containerID="edf0397b7b8e3127ca17cfd8f9a53ae29dbd5aa403b95ff6fbf64049fe7163f1" Mar 18 00:35:32 crc kubenswrapper[4983]: E0318 00:35:32.900027 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edf0397b7b8e3127ca17cfd8f9a53ae29dbd5aa403b95ff6fbf64049fe7163f1\": container with ID starting with edf0397b7b8e3127ca17cfd8f9a53ae29dbd5aa403b95ff6fbf64049fe7163f1 not found: ID does not exist" containerID="edf0397b7b8e3127ca17cfd8f9a53ae29dbd5aa403b95ff6fbf64049fe7163f1" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.900083 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edf0397b7b8e3127ca17cfd8f9a53ae29dbd5aa403b95ff6fbf64049fe7163f1"} err="failed to get container status \"edf0397b7b8e3127ca17cfd8f9a53ae29dbd5aa403b95ff6fbf64049fe7163f1\": rpc error: code = NotFound desc = could not find container \"edf0397b7b8e3127ca17cfd8f9a53ae29dbd5aa403b95ff6fbf64049fe7163f1\": container with ID starting with edf0397b7b8e3127ca17cfd8f9a53ae29dbd5aa403b95ff6fbf64049fe7163f1 not found: ID does not exist" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.900116 4983 scope.go:117] "RemoveContainer" containerID="7230bc0f9e808b6be448dd8cfa4c4203d28240580797cc32851f9de2ebbfff95" Mar 18 00:35:32 crc kubenswrapper[4983]: E0318 00:35:32.900878 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7230bc0f9e808b6be448dd8cfa4c4203d28240580797cc32851f9de2ebbfff95\": container with ID starting with 7230bc0f9e808b6be448dd8cfa4c4203d28240580797cc32851f9de2ebbfff95 not found: ID does not exist" containerID="7230bc0f9e808b6be448dd8cfa4c4203d28240580797cc32851f9de2ebbfff95" Mar 18 00:35:32 crc kubenswrapper[4983]: I0318 00:35:32.900910 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7230bc0f9e808b6be448dd8cfa4c4203d28240580797cc32851f9de2ebbfff95"} err="failed to get container status \"7230bc0f9e808b6be448dd8cfa4c4203d28240580797cc32851f9de2ebbfff95\": rpc error: code = NotFound desc = could not find container \"7230bc0f9e808b6be448dd8cfa4c4203d28240580797cc32851f9de2ebbfff95\": container with ID starting with 7230bc0f9e808b6be448dd8cfa4c4203d28240580797cc32851f9de2ebbfff95 not found: ID does not exist" Mar 18 00:35:34 crc kubenswrapper[4983]: I0318 00:35:34.858270 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35efb54b-6131-47cd-b092-28187491df7f" path="/var/lib/kubelet/pods/35efb54b-6131-47cd-b092-28187491df7f/volumes" Mar 18 00:35:34 crc kubenswrapper[4983]: I0318 00:35:34.965208 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:35 crc kubenswrapper[4983]: I0318 00:35:35.047621 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:35 crc kubenswrapper[4983]: I0318 00:35:35.212994 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lfz8b"] Mar 18 00:35:36 crc kubenswrapper[4983]: I0318 00:35:36.846432 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lfz8b" podUID="2437ad60-b755-4cc9-ad65-084d8b4ca03e" containerName="registry-server" containerID="cri-o://7b3fd9b32e3354a48acbc973efe1276b89c787168b7b97543d79ede73fbe6c25" gracePeriod=2 Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.293235 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.378428 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nxzl\" (UniqueName: \"kubernetes.io/projected/2437ad60-b755-4cc9-ad65-084d8b4ca03e-kube-api-access-2nxzl\") pod \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\" (UID: \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\") " Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.378553 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2437ad60-b755-4cc9-ad65-084d8b4ca03e-catalog-content\") pod \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\" (UID: \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\") " Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.378592 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2437ad60-b755-4cc9-ad65-084d8b4ca03e-utilities\") pod \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\" (UID: \"2437ad60-b755-4cc9-ad65-084d8b4ca03e\") " Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.380635 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2437ad60-b755-4cc9-ad65-084d8b4ca03e-utilities" (OuterVolumeSpecName: "utilities") pod "2437ad60-b755-4cc9-ad65-084d8b4ca03e" (UID: "2437ad60-b755-4cc9-ad65-084d8b4ca03e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.386363 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2437ad60-b755-4cc9-ad65-084d8b4ca03e-kube-api-access-2nxzl" (OuterVolumeSpecName: "kube-api-access-2nxzl") pod "2437ad60-b755-4cc9-ad65-084d8b4ca03e" (UID: "2437ad60-b755-4cc9-ad65-084d8b4ca03e"). InnerVolumeSpecName "kube-api-access-2nxzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.480348 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2437ad60-b755-4cc9-ad65-084d8b4ca03e-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.480508 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nxzl\" (UniqueName: \"kubernetes.io/projected/2437ad60-b755-4cc9-ad65-084d8b4ca03e-kube-api-access-2nxzl\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.512758 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2437ad60-b755-4cc9-ad65-084d8b4ca03e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2437ad60-b755-4cc9-ad65-084d8b4ca03e" (UID: "2437ad60-b755-4cc9-ad65-084d8b4ca03e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.582547 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2437ad60-b755-4cc9-ad65-084d8b4ca03e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.855592 4983 generic.go:334] "Generic (PLEG): container finished" podID="2437ad60-b755-4cc9-ad65-084d8b4ca03e" containerID="7b3fd9b32e3354a48acbc973efe1276b89c787168b7b97543d79ede73fbe6c25" exitCode=0 Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.855639 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfz8b" event={"ID":"2437ad60-b755-4cc9-ad65-084d8b4ca03e","Type":"ContainerDied","Data":"7b3fd9b32e3354a48acbc973efe1276b89c787168b7b97543d79ede73fbe6c25"} Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.855667 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfz8b" event={"ID":"2437ad60-b755-4cc9-ad65-084d8b4ca03e","Type":"ContainerDied","Data":"2f4828f58482c373836d3f3da3f04f59308b714ca6ec25c2ca17a731686ad637"} Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.855686 4983 scope.go:117] "RemoveContainer" containerID="7b3fd9b32e3354a48acbc973efe1276b89c787168b7b97543d79ede73fbe6c25" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.855834 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lfz8b" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.885882 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lfz8b"] Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.891297 4983 scope.go:117] "RemoveContainer" containerID="1376843a0a09d58d678b53577403712502643d21cc36f5a25b475f6f0abd97db" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.897111 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lfz8b"] Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.916055 4983 scope.go:117] "RemoveContainer" containerID="f7bb2c2226d74d71536f9100c7f562c13d7876e0020511233163d67ea55f2f39" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.945023 4983 scope.go:117] "RemoveContainer" containerID="7b3fd9b32e3354a48acbc973efe1276b89c787168b7b97543d79ede73fbe6c25" Mar 18 00:35:37 crc kubenswrapper[4983]: E0318 00:35:37.945573 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b3fd9b32e3354a48acbc973efe1276b89c787168b7b97543d79ede73fbe6c25\": container with ID starting with 7b3fd9b32e3354a48acbc973efe1276b89c787168b7b97543d79ede73fbe6c25 not found: ID does not exist" containerID="7b3fd9b32e3354a48acbc973efe1276b89c787168b7b97543d79ede73fbe6c25" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.945618 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b3fd9b32e3354a48acbc973efe1276b89c787168b7b97543d79ede73fbe6c25"} err="failed to get container status \"7b3fd9b32e3354a48acbc973efe1276b89c787168b7b97543d79ede73fbe6c25\": rpc error: code = NotFound desc = could not find container \"7b3fd9b32e3354a48acbc973efe1276b89c787168b7b97543d79ede73fbe6c25\": container with ID starting with 7b3fd9b32e3354a48acbc973efe1276b89c787168b7b97543d79ede73fbe6c25 not found: ID does not exist" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.945644 4983 scope.go:117] "RemoveContainer" containerID="1376843a0a09d58d678b53577403712502643d21cc36f5a25b475f6f0abd97db" Mar 18 00:35:37 crc kubenswrapper[4983]: E0318 00:35:37.945988 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1376843a0a09d58d678b53577403712502643d21cc36f5a25b475f6f0abd97db\": container with ID starting with 1376843a0a09d58d678b53577403712502643d21cc36f5a25b475f6f0abd97db not found: ID does not exist" containerID="1376843a0a09d58d678b53577403712502643d21cc36f5a25b475f6f0abd97db" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.946026 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1376843a0a09d58d678b53577403712502643d21cc36f5a25b475f6f0abd97db"} err="failed to get container status \"1376843a0a09d58d678b53577403712502643d21cc36f5a25b475f6f0abd97db\": rpc error: code = NotFound desc = could not find container \"1376843a0a09d58d678b53577403712502643d21cc36f5a25b475f6f0abd97db\": container with ID starting with 1376843a0a09d58d678b53577403712502643d21cc36f5a25b475f6f0abd97db not found: ID does not exist" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.946049 4983 scope.go:117] "RemoveContainer" containerID="f7bb2c2226d74d71536f9100c7f562c13d7876e0020511233163d67ea55f2f39" Mar 18 00:35:37 crc kubenswrapper[4983]: E0318 00:35:37.946408 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7bb2c2226d74d71536f9100c7f562c13d7876e0020511233163d67ea55f2f39\": container with ID starting with f7bb2c2226d74d71536f9100c7f562c13d7876e0020511233163d67ea55f2f39 not found: ID does not exist" containerID="f7bb2c2226d74d71536f9100c7f562c13d7876e0020511233163d67ea55f2f39" Mar 18 00:35:37 crc kubenswrapper[4983]: I0318 00:35:37.946435 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7bb2c2226d74d71536f9100c7f562c13d7876e0020511233163d67ea55f2f39"} err="failed to get container status \"f7bb2c2226d74d71536f9100c7f562c13d7876e0020511233163d67ea55f2f39\": rpc error: code = NotFound desc = could not find container \"f7bb2c2226d74d71536f9100c7f562c13d7876e0020511233163d67ea55f2f39\": container with ID starting with f7bb2c2226d74d71536f9100c7f562c13d7876e0020511233163d67ea55f2f39 not found: ID does not exist" Mar 18 00:35:38 crc kubenswrapper[4983]: I0318 00:35:38.856417 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2437ad60-b755-4cc9-ad65-084d8b4ca03e" path="/var/lib/kubelet/pods/2437ad60-b755-4cc9-ad65-084d8b4ca03e/volumes" Mar 18 00:35:39 crc kubenswrapper[4983]: I0318 00:35:39.844360 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:35:39 crc kubenswrapper[4983]: E0318 00:35:39.844795 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.621742 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6rpvl"] Mar 18 00:35:40 crc kubenswrapper[4983]: E0318 00:35:40.622174 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2437ad60-b755-4cc9-ad65-084d8b4ca03e" containerName="extract-content" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.622197 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2437ad60-b755-4cc9-ad65-084d8b4ca03e" containerName="extract-content" Mar 18 00:35:40 crc kubenswrapper[4983]: E0318 00:35:40.622212 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2437ad60-b755-4cc9-ad65-084d8b4ca03e" containerName="registry-server" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.622223 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2437ad60-b755-4cc9-ad65-084d8b4ca03e" containerName="registry-server" Mar 18 00:35:40 crc kubenswrapper[4983]: E0318 00:35:40.622245 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35efb54b-6131-47cd-b092-28187491df7f" containerName="extract-utilities" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.622258 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="35efb54b-6131-47cd-b092-28187491df7f" containerName="extract-utilities" Mar 18 00:35:40 crc kubenswrapper[4983]: E0318 00:35:40.622277 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35efb54b-6131-47cd-b092-28187491df7f" containerName="extract-content" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.622288 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="35efb54b-6131-47cd-b092-28187491df7f" containerName="extract-content" Mar 18 00:35:40 crc kubenswrapper[4983]: E0318 00:35:40.622308 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35efb54b-6131-47cd-b092-28187491df7f" containerName="registry-server" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.622318 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="35efb54b-6131-47cd-b092-28187491df7f" containerName="registry-server" Mar 18 00:35:40 crc kubenswrapper[4983]: E0318 00:35:40.622337 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2437ad60-b755-4cc9-ad65-084d8b4ca03e" containerName="extract-utilities" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.622347 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="2437ad60-b755-4cc9-ad65-084d8b4ca03e" containerName="extract-utilities" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.622504 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="2437ad60-b755-4cc9-ad65-084d8b4ca03e" containerName="registry-server" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.622521 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="35efb54b-6131-47cd-b092-28187491df7f" containerName="registry-server" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.623737 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.642246 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6rpvl"] Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.727681 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nzbt\" (UniqueName: \"kubernetes.io/projected/8eb5923c-c653-4dba-954d-0a6daf775f65-kube-api-access-5nzbt\") pod \"community-operators-6rpvl\" (UID: \"8eb5923c-c653-4dba-954d-0a6daf775f65\") " pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.727766 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eb5923c-c653-4dba-954d-0a6daf775f65-catalog-content\") pod \"community-operators-6rpvl\" (UID: \"8eb5923c-c653-4dba-954d-0a6daf775f65\") " pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.727799 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eb5923c-c653-4dba-954d-0a6daf775f65-utilities\") pod \"community-operators-6rpvl\" (UID: \"8eb5923c-c653-4dba-954d-0a6daf775f65\") " pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.829551 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nzbt\" (UniqueName: \"kubernetes.io/projected/8eb5923c-c653-4dba-954d-0a6daf775f65-kube-api-access-5nzbt\") pod \"community-operators-6rpvl\" (UID: \"8eb5923c-c653-4dba-954d-0a6daf775f65\") " pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.829664 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eb5923c-c653-4dba-954d-0a6daf775f65-catalog-content\") pod \"community-operators-6rpvl\" (UID: \"8eb5923c-c653-4dba-954d-0a6daf775f65\") " pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.829751 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eb5923c-c653-4dba-954d-0a6daf775f65-utilities\") pod \"community-operators-6rpvl\" (UID: \"8eb5923c-c653-4dba-954d-0a6daf775f65\") " pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.830222 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eb5923c-c653-4dba-954d-0a6daf775f65-catalog-content\") pod \"community-operators-6rpvl\" (UID: \"8eb5923c-c653-4dba-954d-0a6daf775f65\") " pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.830376 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eb5923c-c653-4dba-954d-0a6daf775f65-utilities\") pod \"community-operators-6rpvl\" (UID: \"8eb5923c-c653-4dba-954d-0a6daf775f65\") " pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.851551 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nzbt\" (UniqueName: \"kubernetes.io/projected/8eb5923c-c653-4dba-954d-0a6daf775f65-kube-api-access-5nzbt\") pod \"community-operators-6rpvl\" (UID: \"8eb5923c-c653-4dba-954d-0a6daf775f65\") " pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:40 crc kubenswrapper[4983]: I0318 00:35:40.949806 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:41 crc kubenswrapper[4983]: I0318 00:35:41.244016 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6rpvl"] Mar 18 00:35:41 crc kubenswrapper[4983]: I0318 00:35:41.893184 4983 generic.go:334] "Generic (PLEG): container finished" podID="8eb5923c-c653-4dba-954d-0a6daf775f65" containerID="d07a984026c7167321487813c6ab221af27f09db463cc0240e775f267a567f30" exitCode=0 Mar 18 00:35:41 crc kubenswrapper[4983]: I0318 00:35:41.893264 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rpvl" event={"ID":"8eb5923c-c653-4dba-954d-0a6daf775f65","Type":"ContainerDied","Data":"d07a984026c7167321487813c6ab221af27f09db463cc0240e775f267a567f30"} Mar 18 00:35:41 crc kubenswrapper[4983]: I0318 00:35:41.893523 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rpvl" event={"ID":"8eb5923c-c653-4dba-954d-0a6daf775f65","Type":"ContainerStarted","Data":"4a31740c9466a571fc661c8398c4eebd7dd67478fd07f5d68f6dbdb19823f005"} Mar 18 00:35:42 crc kubenswrapper[4983]: I0318 00:35:42.902896 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rpvl" event={"ID":"8eb5923c-c653-4dba-954d-0a6daf775f65","Type":"ContainerStarted","Data":"6d3415b5f6f199d88b9d735a1e9ce50dea5d0de6796664efeb79f57fb1dc9d08"} Mar 18 00:35:43 crc kubenswrapper[4983]: I0318 00:35:43.914401 4983 generic.go:334] "Generic (PLEG): container finished" podID="8eb5923c-c653-4dba-954d-0a6daf775f65" containerID="6d3415b5f6f199d88b9d735a1e9ce50dea5d0de6796664efeb79f57fb1dc9d08" exitCode=0 Mar 18 00:35:43 crc kubenswrapper[4983]: I0318 00:35:43.914446 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rpvl" event={"ID":"8eb5923c-c653-4dba-954d-0a6daf775f65","Type":"ContainerDied","Data":"6d3415b5f6f199d88b9d735a1e9ce50dea5d0de6796664efeb79f57fb1dc9d08"} Mar 18 00:35:44 crc kubenswrapper[4983]: I0318 00:35:44.924902 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rpvl" event={"ID":"8eb5923c-c653-4dba-954d-0a6daf775f65","Type":"ContainerStarted","Data":"9273fddb66f1374b009549791b87a768249f6de7fa10974c4ea0837a9f73c8c2"} Mar 18 00:35:44 crc kubenswrapper[4983]: I0318 00:35:44.952049 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6rpvl" podStartSLOduration=2.534372693 podStartE2EDuration="4.952025636s" podCreationTimestamp="2026-03-18 00:35:40 +0000 UTC" firstStartedPulling="2026-03-18 00:35:41.894854935 +0000 UTC m=+1659.792229575" lastFinishedPulling="2026-03-18 00:35:44.312507868 +0000 UTC m=+1662.209882518" observedRunningTime="2026-03-18 00:35:44.944222687 +0000 UTC m=+1662.841597347" watchObservedRunningTime="2026-03-18 00:35:44.952025636 +0000 UTC m=+1662.849400256" Mar 18 00:35:50 crc kubenswrapper[4983]: I0318 00:35:50.950263 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:50 crc kubenswrapper[4983]: I0318 00:35:50.950861 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:51 crc kubenswrapper[4983]: I0318 00:35:51.029667 4983 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:51 crc kubenswrapper[4983]: I0318 00:35:51.103457 4983 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:51 crc kubenswrapper[4983]: I0318 00:35:51.270910 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6rpvl"] Mar 18 00:35:53 crc kubenswrapper[4983]: I0318 00:35:53.001031 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6rpvl" podUID="8eb5923c-c653-4dba-954d-0a6daf775f65" containerName="registry-server" containerID="cri-o://9273fddb66f1374b009549791b87a768249f6de7fa10974c4ea0837a9f73c8c2" gracePeriod=2 Mar 18 00:35:53 crc kubenswrapper[4983]: I0318 00:35:53.382394 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:53 crc kubenswrapper[4983]: I0318 00:35:53.433891 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eb5923c-c653-4dba-954d-0a6daf775f65-catalog-content\") pod \"8eb5923c-c653-4dba-954d-0a6daf775f65\" (UID: \"8eb5923c-c653-4dba-954d-0a6daf775f65\") " Mar 18 00:35:53 crc kubenswrapper[4983]: I0318 00:35:53.434125 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nzbt\" (UniqueName: \"kubernetes.io/projected/8eb5923c-c653-4dba-954d-0a6daf775f65-kube-api-access-5nzbt\") pod \"8eb5923c-c653-4dba-954d-0a6daf775f65\" (UID: \"8eb5923c-c653-4dba-954d-0a6daf775f65\") " Mar 18 00:35:53 crc kubenswrapper[4983]: I0318 00:35:53.434225 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eb5923c-c653-4dba-954d-0a6daf775f65-utilities\") pod \"8eb5923c-c653-4dba-954d-0a6daf775f65\" (UID: \"8eb5923c-c653-4dba-954d-0a6daf775f65\") " Mar 18 00:35:53 crc kubenswrapper[4983]: I0318 00:35:53.435315 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8eb5923c-c653-4dba-954d-0a6daf775f65-utilities" (OuterVolumeSpecName: "utilities") pod "8eb5923c-c653-4dba-954d-0a6daf775f65" (UID: "8eb5923c-c653-4dba-954d-0a6daf775f65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:35:53 crc kubenswrapper[4983]: I0318 00:35:53.440751 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8eb5923c-c653-4dba-954d-0a6daf775f65-kube-api-access-5nzbt" (OuterVolumeSpecName: "kube-api-access-5nzbt") pod "8eb5923c-c653-4dba-954d-0a6daf775f65" (UID: "8eb5923c-c653-4dba-954d-0a6daf775f65"). InnerVolumeSpecName "kube-api-access-5nzbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:35:53 crc kubenswrapper[4983]: I0318 00:35:53.536670 4983 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eb5923c-c653-4dba-954d-0a6daf775f65-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:53 crc kubenswrapper[4983]: I0318 00:35:53.536719 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nzbt\" (UniqueName: \"kubernetes.io/projected/8eb5923c-c653-4dba-954d-0a6daf775f65-kube-api-access-5nzbt\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:53 crc kubenswrapper[4983]: I0318 00:35:53.935197 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8eb5923c-c653-4dba-954d-0a6daf775f65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8eb5923c-c653-4dba-954d-0a6daf775f65" (UID: "8eb5923c-c653-4dba-954d-0a6daf775f65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:35:53 crc kubenswrapper[4983]: I0318 00:35:53.942007 4983 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eb5923c-c653-4dba-954d-0a6daf775f65-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.011891 4983 generic.go:334] "Generic (PLEG): container finished" podID="8eb5923c-c653-4dba-954d-0a6daf775f65" containerID="9273fddb66f1374b009549791b87a768249f6de7fa10974c4ea0837a9f73c8c2" exitCode=0 Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.011972 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rpvl" Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.011981 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rpvl" event={"ID":"8eb5923c-c653-4dba-954d-0a6daf775f65","Type":"ContainerDied","Data":"9273fddb66f1374b009549791b87a768249f6de7fa10974c4ea0837a9f73c8c2"} Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.013095 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rpvl" event={"ID":"8eb5923c-c653-4dba-954d-0a6daf775f65","Type":"ContainerDied","Data":"4a31740c9466a571fc661c8398c4eebd7dd67478fd07f5d68f6dbdb19823f005"} Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.013171 4983 scope.go:117] "RemoveContainer" containerID="9273fddb66f1374b009549791b87a768249f6de7fa10974c4ea0837a9f73c8c2" Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.043833 4983 scope.go:117] "RemoveContainer" containerID="6d3415b5f6f199d88b9d735a1e9ce50dea5d0de6796664efeb79f57fb1dc9d08" Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.054202 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6rpvl"] Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.064589 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6rpvl"] Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.078857 4983 scope.go:117] "RemoveContainer" containerID="d07a984026c7167321487813c6ab221af27f09db463cc0240e775f267a567f30" Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.120147 4983 scope.go:117] "RemoveContainer" containerID="9273fddb66f1374b009549791b87a768249f6de7fa10974c4ea0837a9f73c8c2" Mar 18 00:35:54 crc kubenswrapper[4983]: E0318 00:35:54.123493 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9273fddb66f1374b009549791b87a768249f6de7fa10974c4ea0837a9f73c8c2\": container with ID starting with 9273fddb66f1374b009549791b87a768249f6de7fa10974c4ea0837a9f73c8c2 not found: ID does not exist" containerID="9273fddb66f1374b009549791b87a768249f6de7fa10974c4ea0837a9f73c8c2" Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.123523 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9273fddb66f1374b009549791b87a768249f6de7fa10974c4ea0837a9f73c8c2"} err="failed to get container status \"9273fddb66f1374b009549791b87a768249f6de7fa10974c4ea0837a9f73c8c2\": rpc error: code = NotFound desc = could not find container \"9273fddb66f1374b009549791b87a768249f6de7fa10974c4ea0837a9f73c8c2\": container with ID starting with 9273fddb66f1374b009549791b87a768249f6de7fa10974c4ea0837a9f73c8c2 not found: ID does not exist" Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.123547 4983 scope.go:117] "RemoveContainer" containerID="6d3415b5f6f199d88b9d735a1e9ce50dea5d0de6796664efeb79f57fb1dc9d08" Mar 18 00:35:54 crc kubenswrapper[4983]: E0318 00:35:54.124828 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d3415b5f6f199d88b9d735a1e9ce50dea5d0de6796664efeb79f57fb1dc9d08\": container with ID starting with 6d3415b5f6f199d88b9d735a1e9ce50dea5d0de6796664efeb79f57fb1dc9d08 not found: ID does not exist" containerID="6d3415b5f6f199d88b9d735a1e9ce50dea5d0de6796664efeb79f57fb1dc9d08" Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.124845 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d3415b5f6f199d88b9d735a1e9ce50dea5d0de6796664efeb79f57fb1dc9d08"} err="failed to get container status \"6d3415b5f6f199d88b9d735a1e9ce50dea5d0de6796664efeb79f57fb1dc9d08\": rpc error: code = NotFound desc = could not find container \"6d3415b5f6f199d88b9d735a1e9ce50dea5d0de6796664efeb79f57fb1dc9d08\": container with ID starting with 6d3415b5f6f199d88b9d735a1e9ce50dea5d0de6796664efeb79f57fb1dc9d08 not found: ID does not exist" Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.124858 4983 scope.go:117] "RemoveContainer" containerID="d07a984026c7167321487813c6ab221af27f09db463cc0240e775f267a567f30" Mar 18 00:35:54 crc kubenswrapper[4983]: E0318 00:35:54.125133 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d07a984026c7167321487813c6ab221af27f09db463cc0240e775f267a567f30\": container with ID starting with d07a984026c7167321487813c6ab221af27f09db463cc0240e775f267a567f30 not found: ID does not exist" containerID="d07a984026c7167321487813c6ab221af27f09db463cc0240e775f267a567f30" Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.125148 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d07a984026c7167321487813c6ab221af27f09db463cc0240e775f267a567f30"} err="failed to get container status \"d07a984026c7167321487813c6ab221af27f09db463cc0240e775f267a567f30\": rpc error: code = NotFound desc = could not find container \"d07a984026c7167321487813c6ab221af27f09db463cc0240e775f267a567f30\": container with ID starting with d07a984026c7167321487813c6ab221af27f09db463cc0240e775f267a567f30 not found: ID does not exist" Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.845234 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:35:54 crc kubenswrapper[4983]: E0318 00:35:54.845921 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:35:54 crc kubenswrapper[4983]: I0318 00:35:54.861060 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8eb5923c-c653-4dba-954d-0a6daf775f65" path="/var/lib/kubelet/pods/8eb5923c-c653-4dba-954d-0a6daf775f65/volumes" Mar 18 00:36:00 crc kubenswrapper[4983]: I0318 00:36:00.700861 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563236-fsfjf"] Mar 18 00:36:00 crc kubenswrapper[4983]: E0318 00:36:00.701747 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eb5923c-c653-4dba-954d-0a6daf775f65" containerName="extract-content" Mar 18 00:36:00 crc kubenswrapper[4983]: I0318 00:36:00.701776 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eb5923c-c653-4dba-954d-0a6daf775f65" containerName="extract-content" Mar 18 00:36:00 crc kubenswrapper[4983]: E0318 00:36:00.701826 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eb5923c-c653-4dba-954d-0a6daf775f65" containerName="extract-utilities" Mar 18 00:36:00 crc kubenswrapper[4983]: I0318 00:36:00.701845 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eb5923c-c653-4dba-954d-0a6daf775f65" containerName="extract-utilities" Mar 18 00:36:00 crc kubenswrapper[4983]: E0318 00:36:00.701875 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eb5923c-c653-4dba-954d-0a6daf775f65" containerName="registry-server" Mar 18 00:36:00 crc kubenswrapper[4983]: I0318 00:36:00.701892 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eb5923c-c653-4dba-954d-0a6daf775f65" containerName="registry-server" Mar 18 00:36:00 crc kubenswrapper[4983]: I0318 00:36:00.702221 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eb5923c-c653-4dba-954d-0a6daf775f65" containerName="registry-server" Mar 18 00:36:00 crc kubenswrapper[4983]: I0318 00:36:00.703123 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563236-fsfjf" Mar 18 00:36:00 crc kubenswrapper[4983]: I0318 00:36:00.707286 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:36:00 crc kubenswrapper[4983]: I0318 00:36:00.707731 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:36:00 crc kubenswrapper[4983]: I0318 00:36:00.714356 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:36:00 crc kubenswrapper[4983]: I0318 00:36:00.721689 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563236-fsfjf"] Mar 18 00:36:00 crc kubenswrapper[4983]: I0318 00:36:00.866529 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq454\" (UniqueName: \"kubernetes.io/projected/1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1-kube-api-access-vq454\") pod \"auto-csr-approver-29563236-fsfjf\" (UID: \"1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1\") " pod="openshift-infra/auto-csr-approver-29563236-fsfjf" Mar 18 00:36:00 crc kubenswrapper[4983]: I0318 00:36:00.969276 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq454\" (UniqueName: \"kubernetes.io/projected/1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1-kube-api-access-vq454\") pod \"auto-csr-approver-29563236-fsfjf\" (UID: \"1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1\") " pod="openshift-infra/auto-csr-approver-29563236-fsfjf" Mar 18 00:36:01 crc kubenswrapper[4983]: I0318 00:36:01.017910 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq454\" (UniqueName: \"kubernetes.io/projected/1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1-kube-api-access-vq454\") pod \"auto-csr-approver-29563236-fsfjf\" (UID: \"1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1\") " pod="openshift-infra/auto-csr-approver-29563236-fsfjf" Mar 18 00:36:01 crc kubenswrapper[4983]: I0318 00:36:01.034351 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563236-fsfjf" Mar 18 00:36:01 crc kubenswrapper[4983]: I0318 00:36:01.484021 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563236-fsfjf"] Mar 18 00:36:01 crc kubenswrapper[4983]: W0318 00:36:01.512775 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c8ed8b4_4ca3_4909_9416_879b9c2bdbd1.slice/crio-6518dc8c76a7c74f44d4fa3b7f347fd6e4220a49be62046da3c319edb0cf1a99 WatchSource:0}: Error finding container 6518dc8c76a7c74f44d4fa3b7f347fd6e4220a49be62046da3c319edb0cf1a99: Status 404 returned error can't find the container with id 6518dc8c76a7c74f44d4fa3b7f347fd6e4220a49be62046da3c319edb0cf1a99 Mar 18 00:36:02 crc kubenswrapper[4983]: I0318 00:36:02.092061 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563236-fsfjf" event={"ID":"1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1","Type":"ContainerStarted","Data":"6518dc8c76a7c74f44d4fa3b7f347fd6e4220a49be62046da3c319edb0cf1a99"} Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.100732 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563236-fsfjf" event={"ID":"1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1","Type":"ContainerStarted","Data":"fb8ebc43967b8097c16a6ba2a0e9468e55cfa7ed7da1e62a14448cf9afce7f28"} Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.115825 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29563236-fsfjf" podStartSLOduration=1.867576049 podStartE2EDuration="3.115808399s" podCreationTimestamp="2026-03-18 00:36:00 +0000 UTC" firstStartedPulling="2026-03-18 00:36:01.517443925 +0000 UTC m=+1679.414818575" lastFinishedPulling="2026-03-18 00:36:02.765676305 +0000 UTC m=+1680.663050925" observedRunningTime="2026-03-18 00:36:03.113988528 +0000 UTC m=+1681.011363178" watchObservedRunningTime="2026-03-18 00:36:03.115808399 +0000 UTC m=+1681.013183009" Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.533164 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-j79wm/must-gather-bxbq7"] Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.534366 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j79wm/must-gather-bxbq7" Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.539809 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-j79wm"/"openshift-service-ca.crt" Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.561278 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-j79wm"/"kube-root-ca.crt" Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.570514 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-j79wm/must-gather-bxbq7"] Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.725047 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/05cb462a-22c9-4d76-abcb-b012b50dc909-must-gather-output\") pod \"must-gather-bxbq7\" (UID: \"05cb462a-22c9-4d76-abcb-b012b50dc909\") " pod="openshift-must-gather-j79wm/must-gather-bxbq7" Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.725145 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb6ff\" (UniqueName: \"kubernetes.io/projected/05cb462a-22c9-4d76-abcb-b012b50dc909-kube-api-access-sb6ff\") pod \"must-gather-bxbq7\" (UID: \"05cb462a-22c9-4d76-abcb-b012b50dc909\") " pod="openshift-must-gather-j79wm/must-gather-bxbq7" Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.827085 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/05cb462a-22c9-4d76-abcb-b012b50dc909-must-gather-output\") pod \"must-gather-bxbq7\" (UID: \"05cb462a-22c9-4d76-abcb-b012b50dc909\") " pod="openshift-must-gather-j79wm/must-gather-bxbq7" Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.827182 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb6ff\" (UniqueName: \"kubernetes.io/projected/05cb462a-22c9-4d76-abcb-b012b50dc909-kube-api-access-sb6ff\") pod \"must-gather-bxbq7\" (UID: \"05cb462a-22c9-4d76-abcb-b012b50dc909\") " pod="openshift-must-gather-j79wm/must-gather-bxbq7" Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.827876 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/05cb462a-22c9-4d76-abcb-b012b50dc909-must-gather-output\") pod \"must-gather-bxbq7\" (UID: \"05cb462a-22c9-4d76-abcb-b012b50dc909\") " pod="openshift-must-gather-j79wm/must-gather-bxbq7" Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.852569 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb6ff\" (UniqueName: \"kubernetes.io/projected/05cb462a-22c9-4d76-abcb-b012b50dc909-kube-api-access-sb6ff\") pod \"must-gather-bxbq7\" (UID: \"05cb462a-22c9-4d76-abcb-b012b50dc909\") " pod="openshift-must-gather-j79wm/must-gather-bxbq7" Mar 18 00:36:03 crc kubenswrapper[4983]: I0318 00:36:03.886571 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j79wm/must-gather-bxbq7" Mar 18 00:36:04 crc kubenswrapper[4983]: I0318 00:36:04.119701 4983 generic.go:334] "Generic (PLEG): container finished" podID="1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1" containerID="fb8ebc43967b8097c16a6ba2a0e9468e55cfa7ed7da1e62a14448cf9afce7f28" exitCode=0 Mar 18 00:36:04 crc kubenswrapper[4983]: I0318 00:36:04.120043 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563236-fsfjf" event={"ID":"1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1","Type":"ContainerDied","Data":"fb8ebc43967b8097c16a6ba2a0e9468e55cfa7ed7da1e62a14448cf9afce7f28"} Mar 18 00:36:04 crc kubenswrapper[4983]: I0318 00:36:04.146426 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-j79wm/must-gather-bxbq7"] Mar 18 00:36:05 crc kubenswrapper[4983]: I0318 00:36:05.142451 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j79wm/must-gather-bxbq7" event={"ID":"05cb462a-22c9-4d76-abcb-b012b50dc909","Type":"ContainerStarted","Data":"952c33f9bb8970a59a278e206f5b686719c2d47572fcd6053b4bb51d5b8ec68e"} Mar 18 00:36:06 crc kubenswrapper[4983]: I0318 00:36:05.516991 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563236-fsfjf" Mar 18 00:36:06 crc kubenswrapper[4983]: I0318 00:36:05.555245 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq454\" (UniqueName: \"kubernetes.io/projected/1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1-kube-api-access-vq454\") pod \"1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1\" (UID: \"1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1\") " Mar 18 00:36:06 crc kubenswrapper[4983]: I0318 00:36:05.567077 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1-kube-api-access-vq454" (OuterVolumeSpecName: "kube-api-access-vq454") pod "1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1" (UID: "1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1"). InnerVolumeSpecName "kube-api-access-vq454". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:36:06 crc kubenswrapper[4983]: I0318 00:36:05.656261 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq454\" (UniqueName: \"kubernetes.io/projected/1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1-kube-api-access-vq454\") on node \"crc\" DevicePath \"\"" Mar 18 00:36:06 crc kubenswrapper[4983]: I0318 00:36:05.938194 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563230-brcnh"] Mar 18 00:36:06 crc kubenswrapper[4983]: I0318 00:36:05.950558 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563230-brcnh"] Mar 18 00:36:06 crc kubenswrapper[4983]: I0318 00:36:06.152742 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563236-fsfjf" event={"ID":"1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1","Type":"ContainerDied","Data":"6518dc8c76a7c74f44d4fa3b7f347fd6e4220a49be62046da3c319edb0cf1a99"} Mar 18 00:36:06 crc kubenswrapper[4983]: I0318 00:36:06.152786 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6518dc8c76a7c74f44d4fa3b7f347fd6e4220a49be62046da3c319edb0cf1a99" Mar 18 00:36:06 crc kubenswrapper[4983]: I0318 00:36:06.152858 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563236-fsfjf" Mar 18 00:36:06 crc kubenswrapper[4983]: I0318 00:36:06.853497 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dff6997b-f408-41b6-8541-4af59267837b" path="/var/lib/kubelet/pods/dff6997b-f408-41b6-8541-4af59267837b/volumes" Mar 18 00:36:09 crc kubenswrapper[4983]: I0318 00:36:09.846532 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:36:09 crc kubenswrapper[4983]: E0318 00:36:09.847291 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:36:10 crc kubenswrapper[4983]: I0318 00:36:10.533435 4983 scope.go:117] "RemoveContainer" containerID="766eeeacad8a1883b60a9c90cc628a1563d63b77977d3aa19a09de35cbe13275" Mar 18 00:36:12 crc kubenswrapper[4983]: I0318 00:36:12.220844 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j79wm/must-gather-bxbq7" event={"ID":"05cb462a-22c9-4d76-abcb-b012b50dc909","Type":"ContainerStarted","Data":"85eca6f117c39f6d8f309ea5c867b79c007349d670b313181d555f3161ae296a"} Mar 18 00:36:12 crc kubenswrapper[4983]: I0318 00:36:12.221599 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j79wm/must-gather-bxbq7" event={"ID":"05cb462a-22c9-4d76-abcb-b012b50dc909","Type":"ContainerStarted","Data":"c8fe0844795797a41f0f4ba11868f7ded8f98c99c27f35df10dcde1c65e6e839"} Mar 18 00:36:12 crc kubenswrapper[4983]: I0318 00:36:12.250428 4983 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-j79wm/must-gather-bxbq7" podStartSLOduration=2.481379627 podStartE2EDuration="9.250399644s" podCreationTimestamp="2026-03-18 00:36:03 +0000 UTC" firstStartedPulling="2026-03-18 00:36:04.150534821 +0000 UTC m=+1682.047909421" lastFinishedPulling="2026-03-18 00:36:10.919554828 +0000 UTC m=+1688.816929438" observedRunningTime="2026-03-18 00:36:12.246536525 +0000 UTC m=+1690.143911135" watchObservedRunningTime="2026-03-18 00:36:12.250399644 +0000 UTC m=+1690.147774254" Mar 18 00:36:22 crc kubenswrapper[4983]: I0318 00:36:22.854177 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:36:22 crc kubenswrapper[4983]: E0318 00:36:22.855252 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:36:33 crc kubenswrapper[4983]: I0318 00:36:33.844408 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:36:33 crc kubenswrapper[4983]: E0318 00:36:33.845439 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:36:45 crc kubenswrapper[4983]: I0318 00:36:45.845149 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:36:45 crc kubenswrapper[4983]: E0318 00:36:45.846402 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:36:56 crc kubenswrapper[4983]: I0318 00:36:56.844075 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:36:56 crc kubenswrapper[4983]: E0318 00:36:56.844852 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:36:59 crc kubenswrapper[4983]: I0318 00:36:59.180291 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-q72ss_a4e42b8f-7120-4b47-b8ed-725640893464/control-plane-machine-set-operator/0.log" Mar 18 00:36:59 crc kubenswrapper[4983]: I0318 00:36:59.353629 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4dn4m_4fbfc372-868f-41d5-81d0-efb8a3e88023/kube-rbac-proxy/0.log" Mar 18 00:36:59 crc kubenswrapper[4983]: I0318 00:36:59.369460 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4dn4m_4fbfc372-868f-41d5-81d0-efb8a3e88023/machine-api-operator/0.log" Mar 18 00:37:10 crc kubenswrapper[4983]: I0318 00:37:10.844231 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:37:10 crc kubenswrapper[4983]: E0318 00:37:10.845272 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:37:14 crc kubenswrapper[4983]: I0318 00:37:14.259269 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-lfdvf_7d75c786-fad6-40d0-85e6-799f383cb4ba/cert-manager-controller/0.log" Mar 18 00:37:14 crc kubenswrapper[4983]: I0318 00:37:14.439626 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-2xxpj_726bff1a-cb5b-495b-aa52-61500dbc5616/cert-manager-cainjector/0.log" Mar 18 00:37:14 crc kubenswrapper[4983]: I0318 00:37:14.482975 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-f7fhd_3f721b9c-8bd6-4e8c-b549-fa5ed88a92b2/cert-manager-webhook/0.log" Mar 18 00:37:22 crc kubenswrapper[4983]: I0318 00:37:22.849723 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:37:22 crc kubenswrapper[4983]: E0318 00:37:22.850408 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:37:29 crc kubenswrapper[4983]: I0318 00:37:29.474152 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-8ff7d675-d2zl4_cc11284f-99b8-4e9c-9553-11116529ffec/prometheus-operator/0.log" Mar 18 00:37:29 crc kubenswrapper[4983]: I0318 00:37:29.511524 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh_1c1f23aa-597c-4649-bec2-649b188836a3/prometheus-operator-admission-webhook/0.log" Mar 18 00:37:29 crc kubenswrapper[4983]: I0318 00:37:29.600146 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg_7cac8d26-fa2e-4c01-b8ae-16070116f5ea/prometheus-operator-admission-webhook/0.log" Mar 18 00:37:29 crc kubenswrapper[4983]: I0318 00:37:29.678619 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-6dd7dd855f-9z5qt_99cbbe02-ca82-4830-a0a6-00bf36e34545/operator/0.log" Mar 18 00:37:29 crc kubenswrapper[4983]: I0318 00:37:29.772867 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-749b9786b8-rccft_42a01360-4470-43f7-a2a5-e3262ba6ab1c/perses-operator/0.log" Mar 18 00:37:37 crc kubenswrapper[4983]: I0318 00:37:37.845022 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:37:37 crc kubenswrapper[4983]: E0318 00:37:37.846034 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:37:44 crc kubenswrapper[4983]: I0318 00:37:44.809157 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4_5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2/util/0.log" Mar 18 00:37:44 crc kubenswrapper[4983]: I0318 00:37:44.991348 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4_5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2/pull/0.log" Mar 18 00:37:45 crc kubenswrapper[4983]: I0318 00:37:45.030701 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4_5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2/util/0.log" Mar 18 00:37:45 crc kubenswrapper[4983]: I0318 00:37:45.031329 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4_5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2/pull/0.log" Mar 18 00:37:45 crc kubenswrapper[4983]: I0318 00:37:45.191153 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4_5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2/extract/0.log" Mar 18 00:37:45 crc kubenswrapper[4983]: I0318 00:37:45.194332 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4_5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2/util/0.log" Mar 18 00:37:45 crc kubenswrapper[4983]: I0318 00:37:45.240649 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fm9cm4_5d28b0cd-21c7-4ed7-ab99-5f8e7a7113a2/pull/0.log" Mar 18 00:37:45 crc kubenswrapper[4983]: I0318 00:37:45.418336 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl_b75a947d-18e9-412c-b3d0-2b5e6dcdfed4/util/0.log" Mar 18 00:37:45 crc kubenswrapper[4983]: I0318 00:37:45.613904 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl_b75a947d-18e9-412c-b3d0-2b5e6dcdfed4/pull/0.log" Mar 18 00:37:45 crc kubenswrapper[4983]: I0318 00:37:45.670077 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl_b75a947d-18e9-412c-b3d0-2b5e6dcdfed4/util/0.log" Mar 18 00:37:45 crc kubenswrapper[4983]: I0318 00:37:45.678467 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl_b75a947d-18e9-412c-b3d0-2b5e6dcdfed4/pull/0.log" Mar 18 00:37:45 crc kubenswrapper[4983]: I0318 00:37:45.841340 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl_b75a947d-18e9-412c-b3d0-2b5e6dcdfed4/util/0.log" Mar 18 00:37:45 crc kubenswrapper[4983]: I0318 00:37:45.858045 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl_b75a947d-18e9-412c-b3d0-2b5e6dcdfed4/pull/0.log" Mar 18 00:37:45 crc kubenswrapper[4983]: I0318 00:37:45.858630 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39eqvppl_b75a947d-18e9-412c-b3d0-2b5e6dcdfed4/extract/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.036679 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx_2a7d9731-da61-4628-a0f2-2c981292584c/util/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.202222 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx_2a7d9731-da61-4628-a0f2-2c981292584c/util/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.227107 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx_2a7d9731-da61-4628-a0f2-2c981292584c/pull/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.303117 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx_2a7d9731-da61-4628-a0f2-2c981292584c/pull/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.392369 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx_2a7d9731-da61-4628-a0f2-2c981292584c/util/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.465870 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx_2a7d9731-da61-4628-a0f2-2c981292584c/pull/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.477741 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5lgfzx_2a7d9731-da61-4628-a0f2-2c981292584c/extract/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.597019 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm_12a26068-f547-44fb-b891-b43a072f3667/util/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.748436 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm_12a26068-f547-44fb-b891-b43a072f3667/pull/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.769588 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm_12a26068-f547-44fb-b891-b43a072f3667/pull/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.789034 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm_12a26068-f547-44fb-b891-b43a072f3667/util/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.914842 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm_12a26068-f547-44fb-b891-b43a072f3667/util/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.936305 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm_12a26068-f547-44fb-b891-b43a072f3667/extract/0.log" Mar 18 00:37:46 crc kubenswrapper[4983]: I0318 00:37:46.941482 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rvpnm_12a26068-f547-44fb-b891-b43a072f3667/pull/0.log" Mar 18 00:37:47 crc kubenswrapper[4983]: I0318 00:37:47.067118 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7scqt_970f8e08-cc08-449e-974e-9ae2f62146f7/extract-utilities/0.log" Mar 18 00:37:47 crc kubenswrapper[4983]: I0318 00:37:47.239493 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7scqt_970f8e08-cc08-449e-974e-9ae2f62146f7/extract-content/0.log" Mar 18 00:37:47 crc kubenswrapper[4983]: I0318 00:37:47.275907 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7scqt_970f8e08-cc08-449e-974e-9ae2f62146f7/extract-content/0.log" Mar 18 00:37:47 crc kubenswrapper[4983]: I0318 00:37:47.297783 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7scqt_970f8e08-cc08-449e-974e-9ae2f62146f7/extract-utilities/0.log" Mar 18 00:37:47 crc kubenswrapper[4983]: I0318 00:37:47.439862 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7scqt_970f8e08-cc08-449e-974e-9ae2f62146f7/extract-utilities/0.log" Mar 18 00:37:47 crc kubenswrapper[4983]: I0318 00:37:47.446512 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7scqt_970f8e08-cc08-449e-974e-9ae2f62146f7/extract-content/0.log" Mar 18 00:37:47 crc kubenswrapper[4983]: I0318 00:37:47.793786 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7scqt_970f8e08-cc08-449e-974e-9ae2f62146f7/registry-server/0.log" Mar 18 00:37:47 crc kubenswrapper[4983]: I0318 00:37:47.891579 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v96xz_9521010b-fae6-4407-8411-585c5a2391dc/extract-utilities/0.log" Mar 18 00:37:48 crc kubenswrapper[4983]: I0318 00:37:48.096528 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v96xz_9521010b-fae6-4407-8411-585c5a2391dc/extract-content/0.log" Mar 18 00:37:48 crc kubenswrapper[4983]: I0318 00:37:48.135591 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v96xz_9521010b-fae6-4407-8411-585c5a2391dc/extract-content/0.log" Mar 18 00:37:48 crc kubenswrapper[4983]: I0318 00:37:48.164584 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v96xz_9521010b-fae6-4407-8411-585c5a2391dc/extract-utilities/0.log" Mar 18 00:37:48 crc kubenswrapper[4983]: I0318 00:37:48.317797 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v96xz_9521010b-fae6-4407-8411-585c5a2391dc/extract-utilities/0.log" Mar 18 00:37:48 crc kubenswrapper[4983]: I0318 00:37:48.338744 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v96xz_9521010b-fae6-4407-8411-585c5a2391dc/extract-content/0.log" Mar 18 00:37:48 crc kubenswrapper[4983]: I0318 00:37:48.543494 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2lhsd_0fa6f497-69b3-4033-babc-85299bfeae4b/marketplace-operator/0.log" Mar 18 00:37:48 crc kubenswrapper[4983]: I0318 00:37:48.597573 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v96xz_9521010b-fae6-4407-8411-585c5a2391dc/registry-server/0.log" Mar 18 00:37:48 crc kubenswrapper[4983]: I0318 00:37:48.642956 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bqvw2_9584343e-c8be-4355-9de2-8ecbd1506da5/extract-utilities/0.log" Mar 18 00:37:48 crc kubenswrapper[4983]: I0318 00:37:48.826683 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bqvw2_9584343e-c8be-4355-9de2-8ecbd1506da5/extract-content/0.log" Mar 18 00:37:48 crc kubenswrapper[4983]: I0318 00:37:48.845142 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:37:48 crc kubenswrapper[4983]: E0318 00:37:48.845363 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:37:48 crc kubenswrapper[4983]: I0318 00:37:48.859035 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bqvw2_9584343e-c8be-4355-9de2-8ecbd1506da5/extract-content/0.log" Mar 18 00:37:48 crc kubenswrapper[4983]: I0318 00:37:48.881335 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bqvw2_9584343e-c8be-4355-9de2-8ecbd1506da5/extract-utilities/0.log" Mar 18 00:37:49 crc kubenswrapper[4983]: I0318 00:37:49.005347 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bqvw2_9584343e-c8be-4355-9de2-8ecbd1506da5/extract-utilities/0.log" Mar 18 00:37:49 crc kubenswrapper[4983]: I0318 00:37:49.024481 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bqvw2_9584343e-c8be-4355-9de2-8ecbd1506da5/extract-content/0.log" Mar 18 00:37:49 crc kubenswrapper[4983]: I0318 00:37:49.235589 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bqvw2_9584343e-c8be-4355-9de2-8ecbd1506da5/registry-server/0.log" Mar 18 00:37:59 crc kubenswrapper[4983]: I0318 00:37:59.845823 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:37:59 crc kubenswrapper[4983]: E0318 00:37:59.847853 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:38:00 crc kubenswrapper[4983]: I0318 00:38:00.154389 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563238-77k2k"] Mar 18 00:38:00 crc kubenswrapper[4983]: E0318 00:38:00.155133 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1" containerName="oc" Mar 18 00:38:00 crc kubenswrapper[4983]: I0318 00:38:00.155160 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1" containerName="oc" Mar 18 00:38:00 crc kubenswrapper[4983]: I0318 00:38:00.155373 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1" containerName="oc" Mar 18 00:38:00 crc kubenswrapper[4983]: I0318 00:38:00.156254 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563238-77k2k" Mar 18 00:38:00 crc kubenswrapper[4983]: I0318 00:38:00.158873 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:38:00 crc kubenswrapper[4983]: I0318 00:38:00.159723 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:38:00 crc kubenswrapper[4983]: I0318 00:38:00.164541 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:38:00 crc kubenswrapper[4983]: I0318 00:38:00.179284 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563238-77k2k"] Mar 18 00:38:00 crc kubenswrapper[4983]: I0318 00:38:00.327258 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnvtc\" (UniqueName: \"kubernetes.io/projected/8cf5f8e2-eb1c-4983-a959-e7f1730274a9-kube-api-access-vnvtc\") pod \"auto-csr-approver-29563238-77k2k\" (UID: \"8cf5f8e2-eb1c-4983-a959-e7f1730274a9\") " pod="openshift-infra/auto-csr-approver-29563238-77k2k" Mar 18 00:38:00 crc kubenswrapper[4983]: I0318 00:38:00.429059 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnvtc\" (UniqueName: \"kubernetes.io/projected/8cf5f8e2-eb1c-4983-a959-e7f1730274a9-kube-api-access-vnvtc\") pod \"auto-csr-approver-29563238-77k2k\" (UID: \"8cf5f8e2-eb1c-4983-a959-e7f1730274a9\") " pod="openshift-infra/auto-csr-approver-29563238-77k2k" Mar 18 00:38:00 crc kubenswrapper[4983]: I0318 00:38:00.451810 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnvtc\" (UniqueName: \"kubernetes.io/projected/8cf5f8e2-eb1c-4983-a959-e7f1730274a9-kube-api-access-vnvtc\") pod \"auto-csr-approver-29563238-77k2k\" (UID: \"8cf5f8e2-eb1c-4983-a959-e7f1730274a9\") " pod="openshift-infra/auto-csr-approver-29563238-77k2k" Mar 18 00:38:00 crc kubenswrapper[4983]: I0318 00:38:00.480687 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563238-77k2k" Mar 18 00:38:00 crc kubenswrapper[4983]: I0318 00:38:00.925871 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563238-77k2k"] Mar 18 00:38:00 crc kubenswrapper[4983]: W0318 00:38:00.944566 4983 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cf5f8e2_eb1c_4983_a959_e7f1730274a9.slice/crio-f75de19d81572c214129eb938fb2afc5445bfb798142c7080bf4685c4236da7b WatchSource:0}: Error finding container f75de19d81572c214129eb938fb2afc5445bfb798142c7080bf4685c4236da7b: Status 404 returned error can't find the container with id f75de19d81572c214129eb938fb2afc5445bfb798142c7080bf4685c4236da7b Mar 18 00:38:01 crc kubenswrapper[4983]: I0318 00:38:01.669808 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563238-77k2k" event={"ID":"8cf5f8e2-eb1c-4983-a959-e7f1730274a9","Type":"ContainerStarted","Data":"f75de19d81572c214129eb938fb2afc5445bfb798142c7080bf4685c4236da7b"} Mar 18 00:38:02 crc kubenswrapper[4983]: I0318 00:38:02.677950 4983 generic.go:334] "Generic (PLEG): container finished" podID="8cf5f8e2-eb1c-4983-a959-e7f1730274a9" containerID="fe2f0716d02d2bb186bb91a4297c32ac745728cca852929d5690c89c8543b151" exitCode=0 Mar 18 00:38:02 crc kubenswrapper[4983]: I0318 00:38:02.678043 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563238-77k2k" event={"ID":"8cf5f8e2-eb1c-4983-a959-e7f1730274a9","Type":"ContainerDied","Data":"fe2f0716d02d2bb186bb91a4297c32ac745728cca852929d5690c89c8543b151"} Mar 18 00:38:03 crc kubenswrapper[4983]: I0318 00:38:03.081909 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-bd569bff7-wc2nh_1c1f23aa-597c-4649-bec2-649b188836a3/prometheus-operator-admission-webhook/0.log" Mar 18 00:38:03 crc kubenswrapper[4983]: I0318 00:38:03.098903 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-bd569bff7-xvxgg_7cac8d26-fa2e-4c01-b8ae-16070116f5ea/prometheus-operator-admission-webhook/0.log" Mar 18 00:38:03 crc kubenswrapper[4983]: I0318 00:38:03.104609 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-8ff7d675-d2zl4_cc11284f-99b8-4e9c-9553-11116529ffec/prometheus-operator/0.log" Mar 18 00:38:03 crc kubenswrapper[4983]: I0318 00:38:03.242525 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-749b9786b8-rccft_42a01360-4470-43f7-a2a5-e3262ba6ab1c/perses-operator/0.log" Mar 18 00:38:03 crc kubenswrapper[4983]: I0318 00:38:03.251869 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-6dd7dd855f-9z5qt_99cbbe02-ca82-4830-a0a6-00bf36e34545/operator/0.log" Mar 18 00:38:03 crc kubenswrapper[4983]: I0318 00:38:03.983256 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563238-77k2k" Mar 18 00:38:04 crc kubenswrapper[4983]: I0318 00:38:04.089376 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnvtc\" (UniqueName: \"kubernetes.io/projected/8cf5f8e2-eb1c-4983-a959-e7f1730274a9-kube-api-access-vnvtc\") pod \"8cf5f8e2-eb1c-4983-a959-e7f1730274a9\" (UID: \"8cf5f8e2-eb1c-4983-a959-e7f1730274a9\") " Mar 18 00:38:04 crc kubenswrapper[4983]: I0318 00:38:04.096297 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cf5f8e2-eb1c-4983-a959-e7f1730274a9-kube-api-access-vnvtc" (OuterVolumeSpecName: "kube-api-access-vnvtc") pod "8cf5f8e2-eb1c-4983-a959-e7f1730274a9" (UID: "8cf5f8e2-eb1c-4983-a959-e7f1730274a9"). InnerVolumeSpecName "kube-api-access-vnvtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:38:04 crc kubenswrapper[4983]: I0318 00:38:04.191147 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnvtc\" (UniqueName: \"kubernetes.io/projected/8cf5f8e2-eb1c-4983-a959-e7f1730274a9-kube-api-access-vnvtc\") on node \"crc\" DevicePath \"\"" Mar 18 00:38:04 crc kubenswrapper[4983]: I0318 00:38:04.695199 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563238-77k2k" event={"ID":"8cf5f8e2-eb1c-4983-a959-e7f1730274a9","Type":"ContainerDied","Data":"f75de19d81572c214129eb938fb2afc5445bfb798142c7080bf4685c4236da7b"} Mar 18 00:38:04 crc kubenswrapper[4983]: I0318 00:38:04.695264 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f75de19d81572c214129eb938fb2afc5445bfb798142c7080bf4685c4236da7b" Mar 18 00:38:04 crc kubenswrapper[4983]: I0318 00:38:04.695580 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563238-77k2k" Mar 18 00:38:05 crc kubenswrapper[4983]: I0318 00:38:05.070997 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563232-jl6dt"] Mar 18 00:38:05 crc kubenswrapper[4983]: I0318 00:38:05.080789 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563232-jl6dt"] Mar 18 00:38:06 crc kubenswrapper[4983]: I0318 00:38:06.855729 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5db34645-4368-4df2-9da6-f3b4b119a4cf" path="/var/lib/kubelet/pods/5db34645-4368-4df2-9da6-f3b4b119a4cf/volumes" Mar 18 00:38:11 crc kubenswrapper[4983]: I0318 00:38:11.033811 4983 scope.go:117] "RemoveContainer" containerID="74f29dd94385cfe5b21027acdab148521bd9bc0f3ef6d55200dcc6e76181fea1" Mar 18 00:38:14 crc kubenswrapper[4983]: I0318 00:38:14.845828 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:38:14 crc kubenswrapper[4983]: E0318 00:38:14.849215 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:38:26 crc kubenswrapper[4983]: I0318 00:38:26.845288 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:38:26 crc kubenswrapper[4983]: E0318 00:38:26.846541 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:38:40 crc kubenswrapper[4983]: I0318 00:38:40.844412 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:38:40 crc kubenswrapper[4983]: E0318 00:38:40.845793 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:38:53 crc kubenswrapper[4983]: I0318 00:38:53.844510 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:38:53 crc kubenswrapper[4983]: E0318 00:38:53.845407 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:38:54 crc kubenswrapper[4983]: I0318 00:38:54.203114 4983 generic.go:334] "Generic (PLEG): container finished" podID="05cb462a-22c9-4d76-abcb-b012b50dc909" containerID="85eca6f117c39f6d8f309ea5c867b79c007349d670b313181d555f3161ae296a" exitCode=0 Mar 18 00:38:54 crc kubenswrapper[4983]: I0318 00:38:54.203184 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j79wm/must-gather-bxbq7" event={"ID":"05cb462a-22c9-4d76-abcb-b012b50dc909","Type":"ContainerDied","Data":"85eca6f117c39f6d8f309ea5c867b79c007349d670b313181d555f3161ae296a"} Mar 18 00:38:54 crc kubenswrapper[4983]: I0318 00:38:54.204050 4983 scope.go:117] "RemoveContainer" containerID="85eca6f117c39f6d8f309ea5c867b79c007349d670b313181d555f3161ae296a" Mar 18 00:38:54 crc kubenswrapper[4983]: I0318 00:38:54.838582 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-j79wm_must-gather-bxbq7_05cb462a-22c9-4d76-abcb-b012b50dc909/gather/0.log" Mar 18 00:39:01 crc kubenswrapper[4983]: I0318 00:39:01.397787 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-j79wm/must-gather-bxbq7"] Mar 18 00:39:01 crc kubenswrapper[4983]: I0318 00:39:01.398531 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-j79wm/must-gather-bxbq7" podUID="05cb462a-22c9-4d76-abcb-b012b50dc909" containerName="copy" containerID="cri-o://c8fe0844795797a41f0f4ba11868f7ded8f98c99c27f35df10dcde1c65e6e839" gracePeriod=2 Mar 18 00:39:01 crc kubenswrapper[4983]: I0318 00:39:01.405558 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-j79wm/must-gather-bxbq7"] Mar 18 00:39:01 crc kubenswrapper[4983]: I0318 00:39:01.741054 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-j79wm_must-gather-bxbq7_05cb462a-22c9-4d76-abcb-b012b50dc909/copy/0.log" Mar 18 00:39:01 crc kubenswrapper[4983]: I0318 00:39:01.741490 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j79wm/must-gather-bxbq7" Mar 18 00:39:01 crc kubenswrapper[4983]: I0318 00:39:01.872089 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6ff\" (UniqueName: \"kubernetes.io/projected/05cb462a-22c9-4d76-abcb-b012b50dc909-kube-api-access-sb6ff\") pod \"05cb462a-22c9-4d76-abcb-b012b50dc909\" (UID: \"05cb462a-22c9-4d76-abcb-b012b50dc909\") " Mar 18 00:39:01 crc kubenswrapper[4983]: I0318 00:39:01.872137 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/05cb462a-22c9-4d76-abcb-b012b50dc909-must-gather-output\") pod \"05cb462a-22c9-4d76-abcb-b012b50dc909\" (UID: \"05cb462a-22c9-4d76-abcb-b012b50dc909\") " Mar 18 00:39:01 crc kubenswrapper[4983]: I0318 00:39:01.878336 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05cb462a-22c9-4d76-abcb-b012b50dc909-kube-api-access-sb6ff" (OuterVolumeSpecName: "kube-api-access-sb6ff") pod "05cb462a-22c9-4d76-abcb-b012b50dc909" (UID: "05cb462a-22c9-4d76-abcb-b012b50dc909"). InnerVolumeSpecName "kube-api-access-sb6ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:39:01 crc kubenswrapper[4983]: I0318 00:39:01.943085 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05cb462a-22c9-4d76-abcb-b012b50dc909-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "05cb462a-22c9-4d76-abcb-b012b50dc909" (UID: "05cb462a-22c9-4d76-abcb-b012b50dc909"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 00:39:01 crc kubenswrapper[4983]: I0318 00:39:01.973882 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6ff\" (UniqueName: \"kubernetes.io/projected/05cb462a-22c9-4d76-abcb-b012b50dc909-kube-api-access-sb6ff\") on node \"crc\" DevicePath \"\"" Mar 18 00:39:01 crc kubenswrapper[4983]: I0318 00:39:01.973952 4983 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/05cb462a-22c9-4d76-abcb-b012b50dc909-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 18 00:39:02 crc kubenswrapper[4983]: I0318 00:39:02.473432 4983 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-j79wm_must-gather-bxbq7_05cb462a-22c9-4d76-abcb-b012b50dc909/copy/0.log" Mar 18 00:39:02 crc kubenswrapper[4983]: I0318 00:39:02.474091 4983 generic.go:334] "Generic (PLEG): container finished" podID="05cb462a-22c9-4d76-abcb-b012b50dc909" containerID="c8fe0844795797a41f0f4ba11868f7ded8f98c99c27f35df10dcde1c65e6e839" exitCode=143 Mar 18 00:39:02 crc kubenswrapper[4983]: I0318 00:39:02.474169 4983 scope.go:117] "RemoveContainer" containerID="c8fe0844795797a41f0f4ba11868f7ded8f98c99c27f35df10dcde1c65e6e839" Mar 18 00:39:02 crc kubenswrapper[4983]: I0318 00:39:02.474398 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j79wm/must-gather-bxbq7" Mar 18 00:39:02 crc kubenswrapper[4983]: I0318 00:39:02.516653 4983 scope.go:117] "RemoveContainer" containerID="85eca6f117c39f6d8f309ea5c867b79c007349d670b313181d555f3161ae296a" Mar 18 00:39:02 crc kubenswrapper[4983]: I0318 00:39:02.579644 4983 scope.go:117] "RemoveContainer" containerID="c8fe0844795797a41f0f4ba11868f7ded8f98c99c27f35df10dcde1c65e6e839" Mar 18 00:39:02 crc kubenswrapper[4983]: E0318 00:39:02.580128 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8fe0844795797a41f0f4ba11868f7ded8f98c99c27f35df10dcde1c65e6e839\": container with ID starting with c8fe0844795797a41f0f4ba11868f7ded8f98c99c27f35df10dcde1c65e6e839 not found: ID does not exist" containerID="c8fe0844795797a41f0f4ba11868f7ded8f98c99c27f35df10dcde1c65e6e839" Mar 18 00:39:02 crc kubenswrapper[4983]: I0318 00:39:02.580158 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8fe0844795797a41f0f4ba11868f7ded8f98c99c27f35df10dcde1c65e6e839"} err="failed to get container status \"c8fe0844795797a41f0f4ba11868f7ded8f98c99c27f35df10dcde1c65e6e839\": rpc error: code = NotFound desc = could not find container \"c8fe0844795797a41f0f4ba11868f7ded8f98c99c27f35df10dcde1c65e6e839\": container with ID starting with c8fe0844795797a41f0f4ba11868f7ded8f98c99c27f35df10dcde1c65e6e839 not found: ID does not exist" Mar 18 00:39:02 crc kubenswrapper[4983]: I0318 00:39:02.580177 4983 scope.go:117] "RemoveContainer" containerID="85eca6f117c39f6d8f309ea5c867b79c007349d670b313181d555f3161ae296a" Mar 18 00:39:02 crc kubenswrapper[4983]: E0318 00:39:02.584217 4983 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85eca6f117c39f6d8f309ea5c867b79c007349d670b313181d555f3161ae296a\": container with ID starting with 85eca6f117c39f6d8f309ea5c867b79c007349d670b313181d555f3161ae296a not found: ID does not exist" containerID="85eca6f117c39f6d8f309ea5c867b79c007349d670b313181d555f3161ae296a" Mar 18 00:39:02 crc kubenswrapper[4983]: I0318 00:39:02.584259 4983 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85eca6f117c39f6d8f309ea5c867b79c007349d670b313181d555f3161ae296a"} err="failed to get container status \"85eca6f117c39f6d8f309ea5c867b79c007349d670b313181d555f3161ae296a\": rpc error: code = NotFound desc = could not find container \"85eca6f117c39f6d8f309ea5c867b79c007349d670b313181d555f3161ae296a\": container with ID starting with 85eca6f117c39f6d8f309ea5c867b79c007349d670b313181d555f3161ae296a not found: ID does not exist" Mar 18 00:39:02 crc kubenswrapper[4983]: I0318 00:39:02.870155 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05cb462a-22c9-4d76-abcb-b012b50dc909" path="/var/lib/kubelet/pods/05cb462a-22c9-4d76-abcb-b012b50dc909/volumes" Mar 18 00:39:06 crc kubenswrapper[4983]: I0318 00:39:06.845052 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:39:06 crc kubenswrapper[4983]: E0318 00:39:06.847522 4983 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jczfs_openshift-machine-config-operator(482e2afe-48be-4ea2-8bda-52849c901b52)\"" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" Mar 18 00:39:19 crc kubenswrapper[4983]: I0318 00:39:19.845403 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" Mar 18 00:39:20 crc kubenswrapper[4983]: I0318 00:39:20.638495 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerStarted","Data":"fc4b5b2b648104d51240d865e44756228c16561371fa9f48f16d61409764ca3f"} Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.154692 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563240-2pnml"] Mar 18 00:40:00 crc kubenswrapper[4983]: E0318 00:40:00.155646 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05cb462a-22c9-4d76-abcb-b012b50dc909" containerName="gather" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.155661 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="05cb462a-22c9-4d76-abcb-b012b50dc909" containerName="gather" Mar 18 00:40:00 crc kubenswrapper[4983]: E0318 00:40:00.155683 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cf5f8e2-eb1c-4983-a959-e7f1730274a9" containerName="oc" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.155692 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cf5f8e2-eb1c-4983-a959-e7f1730274a9" containerName="oc" Mar 18 00:40:00 crc kubenswrapper[4983]: E0318 00:40:00.155705 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05cb462a-22c9-4d76-abcb-b012b50dc909" containerName="copy" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.155713 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="05cb462a-22c9-4d76-abcb-b012b50dc909" containerName="copy" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.155843 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="05cb462a-22c9-4d76-abcb-b012b50dc909" containerName="gather" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.155857 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="05cb462a-22c9-4d76-abcb-b012b50dc909" containerName="copy" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.155876 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cf5f8e2-eb1c-4983-a959-e7f1730274a9" containerName="oc" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.156407 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563240-2pnml" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.159690 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.159813 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.185444 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.218170 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563240-2pnml"] Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.349124 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7875\" (UniqueName: \"kubernetes.io/projected/030ff5f0-cc93-413e-a047-a0f7305225ad-kube-api-access-z7875\") pod \"auto-csr-approver-29563240-2pnml\" (UID: \"030ff5f0-cc93-413e-a047-a0f7305225ad\") " pod="openshift-infra/auto-csr-approver-29563240-2pnml" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.450287 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7875\" (UniqueName: \"kubernetes.io/projected/030ff5f0-cc93-413e-a047-a0f7305225ad-kube-api-access-z7875\") pod \"auto-csr-approver-29563240-2pnml\" (UID: \"030ff5f0-cc93-413e-a047-a0f7305225ad\") " pod="openshift-infra/auto-csr-approver-29563240-2pnml" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.476564 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7875\" (UniqueName: \"kubernetes.io/projected/030ff5f0-cc93-413e-a047-a0f7305225ad-kube-api-access-z7875\") pod \"auto-csr-approver-29563240-2pnml\" (UID: \"030ff5f0-cc93-413e-a047-a0f7305225ad\") " pod="openshift-infra/auto-csr-approver-29563240-2pnml" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.508608 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563240-2pnml" Mar 18 00:40:00 crc kubenswrapper[4983]: I0318 00:40:00.725042 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563240-2pnml"] Mar 18 00:40:01 crc kubenswrapper[4983]: I0318 00:40:01.009631 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563240-2pnml" event={"ID":"030ff5f0-cc93-413e-a047-a0f7305225ad","Type":"ContainerStarted","Data":"0af1d0169ab57a7f4ac8d9db1f072ea020749def0a3557f98fe68e86feb3a4c8"} Mar 18 00:40:03 crc kubenswrapper[4983]: I0318 00:40:03.029345 4983 generic.go:334] "Generic (PLEG): container finished" podID="030ff5f0-cc93-413e-a047-a0f7305225ad" containerID="cedadd0d357bc7e7d52f97a9d40a72bb7c1450961e5a25b99d8e7a151b5d81e6" exitCode=0 Mar 18 00:40:03 crc kubenswrapper[4983]: I0318 00:40:03.029427 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563240-2pnml" event={"ID":"030ff5f0-cc93-413e-a047-a0f7305225ad","Type":"ContainerDied","Data":"cedadd0d357bc7e7d52f97a9d40a72bb7c1450961e5a25b99d8e7a151b5d81e6"} Mar 18 00:40:04 crc kubenswrapper[4983]: I0318 00:40:04.318249 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563240-2pnml" Mar 18 00:40:04 crc kubenswrapper[4983]: I0318 00:40:04.509289 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7875\" (UniqueName: \"kubernetes.io/projected/030ff5f0-cc93-413e-a047-a0f7305225ad-kube-api-access-z7875\") pod \"030ff5f0-cc93-413e-a047-a0f7305225ad\" (UID: \"030ff5f0-cc93-413e-a047-a0f7305225ad\") " Mar 18 00:40:04 crc kubenswrapper[4983]: I0318 00:40:04.520779 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/030ff5f0-cc93-413e-a047-a0f7305225ad-kube-api-access-z7875" (OuterVolumeSpecName: "kube-api-access-z7875") pod "030ff5f0-cc93-413e-a047-a0f7305225ad" (UID: "030ff5f0-cc93-413e-a047-a0f7305225ad"). InnerVolumeSpecName "kube-api-access-z7875". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:40:04 crc kubenswrapper[4983]: I0318 00:40:04.611251 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7875\" (UniqueName: \"kubernetes.io/projected/030ff5f0-cc93-413e-a047-a0f7305225ad-kube-api-access-z7875\") on node \"crc\" DevicePath \"\"" Mar 18 00:40:05 crc kubenswrapper[4983]: I0318 00:40:05.059434 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563240-2pnml" event={"ID":"030ff5f0-cc93-413e-a047-a0f7305225ad","Type":"ContainerDied","Data":"0af1d0169ab57a7f4ac8d9db1f072ea020749def0a3557f98fe68e86feb3a4c8"} Mar 18 00:40:05 crc kubenswrapper[4983]: I0318 00:40:05.059661 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0af1d0169ab57a7f4ac8d9db1f072ea020749def0a3557f98fe68e86feb3a4c8" Mar 18 00:40:05 crc kubenswrapper[4983]: I0318 00:40:05.059503 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563240-2pnml" Mar 18 00:40:05 crc kubenswrapper[4983]: I0318 00:40:05.381907 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563234-pzlvr"] Mar 18 00:40:05 crc kubenswrapper[4983]: I0318 00:40:05.386886 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563234-pzlvr"] Mar 18 00:40:06 crc kubenswrapper[4983]: I0318 00:40:06.856561 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f02470c-1df4-4c79-b670-460dcecb6c68" path="/var/lib/kubelet/pods/5f02470c-1df4-4c79-b670-460dcecb6c68/volumes" Mar 18 00:40:11 crc kubenswrapper[4983]: I0318 00:40:11.134126 4983 scope.go:117] "RemoveContainer" containerID="bfbd7f055f41690d2395af1912ddb98be99cfbf2d8fdf33ff26d4d7653e8e8fa" Mar 18 00:41:43 crc kubenswrapper[4983]: I0318 00:41:43.259193 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:41:43 crc kubenswrapper[4983]: I0318 00:41:43.260089 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:42:00 crc kubenswrapper[4983]: I0318 00:42:00.680538 4983 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29563242-z85zv"] Mar 18 00:42:00 crc kubenswrapper[4983]: E0318 00:42:00.681509 4983 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="030ff5f0-cc93-413e-a047-a0f7305225ad" containerName="oc" Mar 18 00:42:00 crc kubenswrapper[4983]: I0318 00:42:00.681529 4983 state_mem.go:107] "Deleted CPUSet assignment" podUID="030ff5f0-cc93-413e-a047-a0f7305225ad" containerName="oc" Mar 18 00:42:00 crc kubenswrapper[4983]: I0318 00:42:00.681723 4983 memory_manager.go:354] "RemoveStaleState removing state" podUID="030ff5f0-cc93-413e-a047-a0f7305225ad" containerName="oc" Mar 18 00:42:00 crc kubenswrapper[4983]: I0318 00:42:00.683070 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563242-z85zv" Mar 18 00:42:00 crc kubenswrapper[4983]: I0318 00:42:00.689371 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 00:42:00 crc kubenswrapper[4983]: I0318 00:42:00.689408 4983 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-qfl56" Mar 18 00:42:00 crc kubenswrapper[4983]: I0318 00:42:00.689546 4983 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 00:42:00 crc kubenswrapper[4983]: I0318 00:42:00.730135 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563242-z85zv"] Mar 18 00:42:00 crc kubenswrapper[4983]: I0318 00:42:00.830270 4983 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w57c\" (UniqueName: \"kubernetes.io/projected/f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef-kube-api-access-5w57c\") pod \"auto-csr-approver-29563242-z85zv\" (UID: \"f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef\") " pod="openshift-infra/auto-csr-approver-29563242-z85zv" Mar 18 00:42:00 crc kubenswrapper[4983]: I0318 00:42:00.931610 4983 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w57c\" (UniqueName: \"kubernetes.io/projected/f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef-kube-api-access-5w57c\") pod \"auto-csr-approver-29563242-z85zv\" (UID: \"f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef\") " pod="openshift-infra/auto-csr-approver-29563242-z85zv" Mar 18 00:42:00 crc kubenswrapper[4983]: I0318 00:42:00.967791 4983 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w57c\" (UniqueName: \"kubernetes.io/projected/f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef-kube-api-access-5w57c\") pod \"auto-csr-approver-29563242-z85zv\" (UID: \"f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef\") " pod="openshift-infra/auto-csr-approver-29563242-z85zv" Mar 18 00:42:01 crc kubenswrapper[4983]: I0318 00:42:01.008660 4983 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563242-z85zv" Mar 18 00:42:01 crc kubenswrapper[4983]: I0318 00:42:01.327764 4983 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29563242-z85zv"] Mar 18 00:42:01 crc kubenswrapper[4983]: I0318 00:42:01.370867 4983 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 00:42:01 crc kubenswrapper[4983]: I0318 00:42:01.670072 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563242-z85zv" event={"ID":"f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef","Type":"ContainerStarted","Data":"67f2a400bf07d410a8efb7e9cd8097ec29d51cd7fdf51d9164929c40aa1d4d67"} Mar 18 00:42:03 crc kubenswrapper[4983]: I0318 00:42:03.689268 4983 generic.go:334] "Generic (PLEG): container finished" podID="f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef" containerID="fd1f1139be9bc3cdec78e1a5a06cff246766a5944172c0573b6c190abaaecf92" exitCode=0 Mar 18 00:42:03 crc kubenswrapper[4983]: I0318 00:42:03.690438 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563242-z85zv" event={"ID":"f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef","Type":"ContainerDied","Data":"fd1f1139be9bc3cdec78e1a5a06cff246766a5944172c0573b6c190abaaecf92"} Mar 18 00:42:05 crc kubenswrapper[4983]: I0318 00:42:05.043276 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563242-z85zv" Mar 18 00:42:05 crc kubenswrapper[4983]: I0318 00:42:05.110715 4983 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w57c\" (UniqueName: \"kubernetes.io/projected/f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef-kube-api-access-5w57c\") pod \"f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef\" (UID: \"f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef\") " Mar 18 00:42:05 crc kubenswrapper[4983]: I0318 00:42:05.122628 4983 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef-kube-api-access-5w57c" (OuterVolumeSpecName: "kube-api-access-5w57c") pod "f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef" (UID: "f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef"). InnerVolumeSpecName "kube-api-access-5w57c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 00:42:05 crc kubenswrapper[4983]: I0318 00:42:05.212219 4983 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w57c\" (UniqueName: \"kubernetes.io/projected/f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef-kube-api-access-5w57c\") on node \"crc\" DevicePath \"\"" Mar 18 00:42:05 crc kubenswrapper[4983]: I0318 00:42:05.711562 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29563242-z85zv" event={"ID":"f99cc7fa-bbb7-4b0f-aeb3-27e36a7069ef","Type":"ContainerDied","Data":"67f2a400bf07d410a8efb7e9cd8097ec29d51cd7fdf51d9164929c40aa1d4d67"} Mar 18 00:42:05 crc kubenswrapper[4983]: I0318 00:42:05.711652 4983 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67f2a400bf07d410a8efb7e9cd8097ec29d51cd7fdf51d9164929c40aa1d4d67" Mar 18 00:42:05 crc kubenswrapper[4983]: I0318 00:42:05.711611 4983 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29563242-z85zv" Mar 18 00:42:06 crc kubenswrapper[4983]: I0318 00:42:06.114287 4983 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29563236-fsfjf"] Mar 18 00:42:06 crc kubenswrapper[4983]: I0318 00:42:06.121037 4983 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29563236-fsfjf"] Mar 18 00:42:06 crc kubenswrapper[4983]: I0318 00:42:06.859199 4983 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1" path="/var/lib/kubelet/pods/1c8ed8b4-4ca3-4909-9416-879b9c2bdbd1/volumes" Mar 18 00:42:11 crc kubenswrapper[4983]: I0318 00:42:11.232065 4983 scope.go:117] "RemoveContainer" containerID="fb8ebc43967b8097c16a6ba2a0e9468e55cfa7ed7da1e62a14448cf9afce7f28" Mar 18 00:42:13 crc kubenswrapper[4983]: I0318 00:42:13.260680 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:42:13 crc kubenswrapper[4983]: I0318 00:42:13.261144 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:42:43 crc kubenswrapper[4983]: I0318 00:42:43.259366 4983 patch_prober.go:28] interesting pod/machine-config-daemon-jczfs container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 00:42:43 crc kubenswrapper[4983]: I0318 00:42:43.260187 4983 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 00:42:43 crc kubenswrapper[4983]: I0318 00:42:43.260271 4983 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" Mar 18 00:42:43 crc kubenswrapper[4983]: I0318 00:42:43.261226 4983 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc4b5b2b648104d51240d865e44756228c16561371fa9f48f16d61409764ca3f"} pod="openshift-machine-config-operator/machine-config-daemon-jczfs" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 00:42:43 crc kubenswrapper[4983]: I0318 00:42:43.261327 4983 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" podUID="482e2afe-48be-4ea2-8bda-52849c901b52" containerName="machine-config-daemon" containerID="cri-o://fc4b5b2b648104d51240d865e44756228c16561371fa9f48f16d61409764ca3f" gracePeriod=600 Mar 18 00:42:44 crc kubenswrapper[4983]: I0318 00:42:44.061985 4983 generic.go:334] "Generic (PLEG): container finished" podID="482e2afe-48be-4ea2-8bda-52849c901b52" containerID="fc4b5b2b648104d51240d865e44756228c16561371fa9f48f16d61409764ca3f" exitCode=0 Mar 18 00:42:44 crc kubenswrapper[4983]: I0318 00:42:44.062235 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerDied","Data":"fc4b5b2b648104d51240d865e44756228c16561371fa9f48f16d61409764ca3f"} Mar 18 00:42:44 crc kubenswrapper[4983]: I0318 00:42:44.062709 4983 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jczfs" event={"ID":"482e2afe-48be-4ea2-8bda-52849c901b52","Type":"ContainerStarted","Data":"3a67b2f19a073d925a37eec3d34eaa5a3a02830bbc059b05f4c9d9be20786c73"} Mar 18 00:42:44 crc kubenswrapper[4983]: I0318 00:42:44.062765 4983 scope.go:117] "RemoveContainer" containerID="fdc85f7c134fa6f778c1d8929059f4c3b1bbbe8de735850d1357bbd80923ddc9" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515156372440024454 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015156372441017372 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015156366000016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015156366001015460 5ustar corecore